[ 509.234991] env[61947]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61947) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 509.235355] env[61947]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61947) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 509.235413] env[61947]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61947) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 509.235713] env[61947]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 509.333782] env[61947]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61947) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 509.343623] env[61947]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61947) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 509.943956] env[61947]: INFO nova.virt.driver [None req-8b71ccbe-2563-4b23-a3e4-3e728c88e651 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 510.014169] env[61947]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.014388] env[61947]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.014454] env[61947]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61947) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 513.104823] env[61947]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-f5dc0d11-1d49-4987-acde-3adca90c422b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.120677] env[61947]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61947) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 513.120814] env[61947]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-20dbee79-80f2-4973-b246-31eb38a42867 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.152998] env[61947]: INFO oslo_vmware.api [-] Successfully established new session; session ID is d49e4. [ 513.153240] env[61947]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.139s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.153693] env[61947]: INFO nova.virt.vmwareapi.driver [None req-8b71ccbe-2563-4b23-a3e4-3e728c88e651 None None] VMware vCenter version: 7.0.3 [ 513.157283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbca8f3-7843-46e1-9452-bb6e74ca9f97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.178810] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc537748-755a-45db-bea8-70912d25c6ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.186334] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613ef412-e442-4113-a0cc-234f03e4c145 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.193542] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230c5884-457e-4adf-948b-cdabfbf350a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.207190] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634e2545-0b49-4a43-9f00-fb74a760c2fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.215104] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d92ac40-71c2-455c-aa79-0363454f02aa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.245848] env[61947]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-cfa62002-1b97-4c5e-9b68-a9ca49df2894 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.252116] env[61947]: DEBUG nova.virt.vmwareapi.driver [None req-8b71ccbe-2563-4b23-a3e4-3e728c88e651 None None] Extension org.openstack.compute already exists. {{(pid=61947) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 513.254969] env[61947]: INFO nova.compute.provider_config [None req-8b71ccbe-2563-4b23-a3e4-3e728c88e651 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 513.758562] env[61947]: DEBUG nova.context [None req-8b71ccbe-2563-4b23-a3e4-3e728c88e651 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),1743c62b-87c6-457c-82bc-82d5c1834096(cell1) {{(pid=61947) load_cells /opt/stack/nova/nova/context.py:464}} [ 513.761670] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.761922] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.762577] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.763024] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Acquiring lock "1743c62b-87c6-457c-82bc-82d5c1834096" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.763236] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Lock "1743c62b-87c6-457c-82bc-82d5c1834096" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.764324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Lock "1743c62b-87c6-457c-82bc-82d5c1834096" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.785380] env[61947]: INFO dbcounter [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Registered counter for database nova_cell0 [ 513.795253] env[61947]: INFO dbcounter [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Registered counter for database nova_cell1 [ 513.798812] env[61947]: DEBUG oslo_db.sqlalchemy.engines [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61947) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 513.799235] env[61947]: DEBUG oslo_db.sqlalchemy.engines [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61947) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 513.805229] env[61947]: ERROR nova.db.main.api [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.805229] env[61947]: result = function(*args, **kwargs) [ 513.805229] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 513.805229] env[61947]: return func(*args, **kwargs) [ 513.805229] env[61947]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 513.805229] env[61947]: result = fn(*args, **kwargs) [ 513.805229] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 513.805229] env[61947]: return f(*args, **kwargs) [ 513.805229] env[61947]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 513.805229] env[61947]: return db.service_get_minimum_version(context, binaries) [ 513.805229] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 513.805229] env[61947]: _check_db_access() [ 513.805229] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 513.805229] env[61947]: stacktrace = ''.join(traceback.format_stack()) [ 513.805229] env[61947]: [ 513.806059] env[61947]: ERROR nova.db.main.api [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.806059] env[61947]: result = function(*args, **kwargs) [ 513.806059] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 513.806059] env[61947]: return func(*args, **kwargs) [ 513.806059] env[61947]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 513.806059] env[61947]: result = fn(*args, **kwargs) [ 513.806059] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 513.806059] env[61947]: return f(*args, **kwargs) [ 513.806059] env[61947]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 513.806059] env[61947]: return db.service_get_minimum_version(context, binaries) [ 513.806059] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 513.806059] env[61947]: _check_db_access() [ 513.806059] env[61947]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 513.806059] env[61947]: stacktrace = ''.join(traceback.format_stack()) [ 513.806059] env[61947]: [ 513.806524] env[61947]: WARNING nova.objects.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Failed to get minimum service version for cell 1743c62b-87c6-457c-82bc-82d5c1834096 [ 513.806779] env[61947]: WARNING nova.objects.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 513.807204] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Acquiring lock "singleton_lock" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.807400] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Acquired lock "singleton_lock" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.808243] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Releasing lock "singleton_lock" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.808621] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Full set of CONF: {{(pid=61947) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 513.808797] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ******************************************************************************** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 513.808943] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Configuration options gathered from: {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 513.809115] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 513.809347] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 513.809505] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ================================================================================ {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 513.809889] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] allow_resize_to_same_host = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.809954] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] arq_binding_timeout = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.810143] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] backdoor_port = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.810315] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] backdoor_socket = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.810517] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] block_device_allocate_retries = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.810712] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] block_device_allocate_retries_interval = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.810907] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cert = self.pem {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.811170] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.811339] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute_monitors = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.811530] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] config_dir = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.811704] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] config_drive_format = iso9660 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.811839] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812019] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] config_source = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812202] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] console_host = devstack {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812367] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] control_exchange = nova {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812563] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cpu_allocation_ratio = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812680] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] daemon = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.812907] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] debug = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.813095] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_access_ip_network_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.813266] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_availability_zone = nova {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.813423] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_ephemeral_format = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.813582] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_green_pool_size = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.813835] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814042] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] default_schedule_zone = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814214] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] disk_allocation_ratio = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814381] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] enable_new_services = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814557] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] enabled_apis = ['osapi_compute'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814724] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] enabled_ssl_apis = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.814887] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] flat_injected = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815057] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] force_config_drive = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815224] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] force_raw_images = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815395] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] graceful_shutdown_timeout = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815557] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] heal_instance_info_cache_interval = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815776] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] host = cpu-1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.815952] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.816137] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.816301] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.816515] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.816680] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_build_timeout = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.816838] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_delete_interval = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817037] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_format = [instance: %(uuid)s] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817259] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_name_template = instance-%08x {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817434] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_usage_audit = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817606] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_usage_audit_period = month {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817772] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.817959] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.818297] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] internal_service_availability_zone = internal {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.818467] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] key = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.818631] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] live_migration_retry_count = 30 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.818801] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_color = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.818966] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_config_append = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819156] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819315] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_dir = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819475] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819607] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_options = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819769] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_rotate_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.819937] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_rotate_interval_type = days {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820118] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] log_rotation_type = none {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820250] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820380] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820546] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820709] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820836] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.820995] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] long_rpc_timeout = 1800 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821173] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_concurrent_builds = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821331] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_concurrent_live_migrations = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821490] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_concurrent_snapshots = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821649] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_local_block_devices = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821808] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_logfile_count = 30 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.821965] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] max_logfile_size_mb = 200 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822139] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] maximum_instance_delete_attempts = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822306] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metadata_listen = 0.0.0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822475] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metadata_listen_port = 8775 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822701] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metadata_workers = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822814] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] migrate_max_retries = -1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.822965] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] mkisofs_cmd = genisoimage {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823191] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823323] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] my_ip = 10.180.1.21 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823486] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] network_allocate_retries = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823662] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823828] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.823989] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] osapi_compute_listen_port = 8774 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824176] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] osapi_compute_unique_server_name_scope = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824343] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] osapi_compute_workers = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824505] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] password_length = 12 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824665] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] periodic_enable = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824824] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] periodic_fuzzy_delay = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.824992] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] pointer_model = usbtablet {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825176] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] preallocate_images = none {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825335] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] publish_errors = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825526] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] pybasedir = /opt/stack/nova {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825622] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ram_allocation_ratio = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825781] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rate_limit_burst = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.825945] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rate_limit_except_level = CRITICAL {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826119] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rate_limit_interval = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826278] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reboot_timeout = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826436] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reclaim_instance_interval = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826591] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] record = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826756] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reimage_timeout_per_gb = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.826947] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] report_interval = 120 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827127] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rescue_timeout = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827291] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reserved_host_cpus = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827450] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reserved_host_disk_mb = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827607] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reserved_host_memory_mb = 512 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827770] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] reserved_huge_pages = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.827938] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] resize_confirm_window = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.828243] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] resize_fs_using_block_device = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.828447] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] resume_guests_state_on_host_boot = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.828625] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.828793] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] rpc_response_timeout = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.828954] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] run_external_periodic_tasks = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829142] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] running_deleted_instance_action = reap {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829310] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829464] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] running_deleted_instance_timeout = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829622] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler_instance_sync_interval = 120 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829790] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_down_time = 720 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.829954] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] servicegroup_driver = db {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830125] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] shell_completion = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830284] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] shelved_offload_time = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830443] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] shelved_poll_interval = 3600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830609] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] shutdown_timeout = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830767] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] source_is_ipv6 = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.830924] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ssl_only = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.831192] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.831363] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] sync_power_state_interval = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.831523] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] sync_power_state_pool_size = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.831690] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] syslog_log_facility = LOG_USER {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.831846] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] tempdir = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832011] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] timeout_nbd = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832190] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] transport_url = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832351] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] update_resources_interval = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832510] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_cow_images = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832667] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_eventlog = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832825] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_journal = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.832980] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_json = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833152] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_rootwrap_daemon = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833308] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_stderr = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833463] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] use_syslog = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833617] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vcpu_pin_set = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833783] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plugging_is_fatal = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.833948] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plugging_timeout = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.834125] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] virt_mkfs = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.834288] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] volume_usage_poll_interval = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.834446] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] watch_log_file = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.834610] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] web = /usr/share/spice-html5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 513.834801] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_concurrency.disable_process_locking = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835110] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835295] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835461] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835629] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835798] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.835963] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.836166] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.auth_strategy = keystone {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.836336] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.compute_link_prefix = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.836511] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.836685] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.dhcp_domain = novalocal {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.836855] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.enable_instance_password = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.837063] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.glance_link_prefix = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.837247] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.837421] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.837585] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.instance_list_per_project_cells = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.837814] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.list_records_by_skipping_down_cells = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.838070] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.local_metadata_per_cell = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.838274] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.max_limit = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.838576] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.metadata_cache_expiration = 15 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.838760] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.neutron_default_tenant_id = default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.838934] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.response_validation = warn {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839124] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.use_neutron_default_nets = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839298] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839463] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839634] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839806] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.839981] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_dynamic_targets = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.840193] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_jsonfile_path = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.840389] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.840588] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.backend = dogpile.cache.memcached {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.840756] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.backend_argument = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.840929] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.config_prefix = cache.oslo {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841117] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.dead_timeout = 60.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841286] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.debug_cache_backend = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841449] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.enable_retry_client = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841610] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.enable_socket_keepalive = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841778] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.enabled = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.841943] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.enforce_fips_mode = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842118] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.expiration_time = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842284] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.hashclient_retry_attempts = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842452] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842612] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_dead_retry = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842771] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_password = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.842930] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843104] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843271] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_pool_maxsize = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843434] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843593] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_sasl_enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843771] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.843939] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.844115] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.memcache_username = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.844300] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.proxies = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.844468] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_db = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.844629] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_password = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.844802] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845010] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845212] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_server = localhost:6379 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845385] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_socket_timeout = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845550] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.redis_username = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845715] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.retry_attempts = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.845887] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.retry_delay = 0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846072] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.socket_keepalive_count = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846246] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.socket_keepalive_idle = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846412] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.socket_keepalive_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846572] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.tls_allowed_ciphers = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846733] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.tls_cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.846934] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.tls_certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.847102] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.tls_enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.847276] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cache.tls_keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.847453] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.847629] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.auth_type = password {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.847791] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.848015] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.848198] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.848365] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.848666] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.cross_az_attach = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.848852] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.debug = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849024] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.endpoint_template = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849203] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.http_retries = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849367] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849526] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849703] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.os_region_name = RegionOne {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.849872] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850046] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cinder.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850240] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850400] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.cpu_dedicated_set = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850559] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.cpu_shared_set = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850726] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.image_type_exclude_list = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.850891] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.851070] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.851238] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.851480] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.851754] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.851954] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.resource_provider_association_refresh = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.852155] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.852340] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.shutdown_retry_interval = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.852532] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.852717] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] conductor.workers = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.852903] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] console.allowed_origins = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853078] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] console.ssl_ciphers = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853255] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] console.ssl_minimum_version = default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853428] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] consoleauth.enforce_session_timeout = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853598] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] consoleauth.token_ttl = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853773] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.853929] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854110] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854274] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854433] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854592] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854755] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.854958] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855141] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855306] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855470] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855636] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855794] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.855967] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.service_type = accelerator {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.856181] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.856351] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.856517] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.856677] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.856863] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857070] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] cyborg.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857265] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.backend = sqlalchemy {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857440] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.connection = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857606] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.connection_debug = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857777] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.connection_parameters = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.857966] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.connection_recycle_time = 3600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.858170] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.connection_trace = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.858339] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.db_inc_retry_interval = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.858503] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.db_max_retries = 20 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.858804] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.db_max_retry_interval = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.858987] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.db_retry_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859174] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.max_overflow = 50 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859340] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.max_pool_size = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859502] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.max_retries = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859673] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859836] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.mysql_wsrep_sync_wait = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.859998] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.pool_timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.860182] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.retry_interval = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.860343] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.slave_connection = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.860505] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.sqlite_synchronous = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.860666] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] database.use_db_reconnect = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.860846] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.backend = sqlalchemy {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861033] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.connection = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861210] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.connection_debug = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861383] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.connection_parameters = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861550] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.connection_recycle_time = 3600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861716] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.connection_trace = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.861881] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.db_inc_retry_interval = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862056] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.db_max_retries = 20 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862228] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.db_max_retry_interval = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862391] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.db_retry_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862551] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.max_overflow = 50 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862713] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.max_pool_size = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.862878] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.max_retries = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863053] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863218] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863378] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.pool_timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863541] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.retry_interval = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863701] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.slave_connection = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.863865] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] api_database.sqlite_synchronous = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864056] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] devices.enabled_mdev_types = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864241] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864415] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864579] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ephemeral_storage_encryption.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864745] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.864921] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.api_servers = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865101] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865267] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865433] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865596] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865759] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.865925] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.debug = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866108] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.default_trusted_certificate_ids = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866281] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.enable_certificate_validation = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866446] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.enable_rbd_download = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866605] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866771] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.866959] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.867159] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.867324] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.867491] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.num_retries = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.867663] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.rbd_ceph_conf = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.867828] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.rbd_connect_timeout = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.868043] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.rbd_pool = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.868230] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.rbd_user = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.868396] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.868555] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.868714] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869049] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.service_type = image {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869233] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869401] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869566] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869727] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.869926] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.870112] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.verify_glance_signatures = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.870278] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] glance.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.870454] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] guestfs.debug = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.870622] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] mks.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871015] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871229] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.manager_interval = 2400 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871405] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.precache_concurrency = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871580] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.remove_unused_base_images = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871753] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.871924] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872121] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] image_cache.subdirectory_name = _base {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872310] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.api_max_retries = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872484] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.api_retry_interval = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872649] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872822] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.auth_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.872987] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873168] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873351] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873508] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.conductor_group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873670] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873833] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.873994] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874180] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874343] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874503] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874662] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874835] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.peer_list = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.874999] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.875180] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.875349] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.serial_console_state_timeout = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.875511] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.875686] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.service_type = baremetal {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.875849] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.shard = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876024] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876194] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876362] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876524] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876712] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.876879] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ironic.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.877116] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.877305] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] key_manager.fixed_key = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.877492] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.877656] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.barbican_api_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.877816] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.barbican_endpoint = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878027] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.barbican_endpoint_type = public {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878207] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.barbican_region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878373] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878535] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878701] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.878864] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.879201] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.879386] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.number_of_retries = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.879560] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.retry_delay = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.879727] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.send_service_user_token = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.879895] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880076] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880249] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.verify_ssl = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880415] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican.verify_ssl_path = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880591] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880757] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.auth_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.880920] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881136] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881317] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881487] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881653] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881820] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.881983] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] barbican_service_user.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.882181] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.approle_role_id = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.882347] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.approle_secret_id = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.882526] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.kv_mountpoint = secret {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.882690] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.kv_path = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.882859] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.kv_version = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883045] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.namespace = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883220] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.root_token_id = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883479] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.ssl_ca_crt_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883554] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.timeout = 60.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883718] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.use_ssl = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.883891] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884113] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884299] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.auth_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884464] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884626] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884792] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.884958] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885141] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885304] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885467] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885626] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885790] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.885951] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886131] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886298] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886460] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886637] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.service_type = identity {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886805] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.886990] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.887242] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.887414] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.887609] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.887775] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] keystone.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.887985] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.connection_uri = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.888167] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_mode = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.888341] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.888513] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_models = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.888687] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_power_governor_high = performance {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.888858] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.889144] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_power_management = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.889367] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.889544] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.device_detach_attempts = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.889712] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.device_detach_timeout = 20 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.889882] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.disk_cachemodes = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890074] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.disk_prefix = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890272] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.enabled_perf_events = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890443] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.file_backed_memory = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890614] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.gid_maps = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890775] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.hw_disk_discard = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.890936] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.hw_machine_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891129] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_rbd_ceph_conf = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891301] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891466] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891643] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_rbd_glance_store_name = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891820] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_rbd_pool = rbd {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.891996] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_type = default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892179] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.images_volume_group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892349] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.inject_key = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892514] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.inject_partition = -2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892841] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.inject_password = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892841] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.iscsi_iface = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.892996] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.iser_use_multipath = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.893205] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.893386] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.893616] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_downtime = 500 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.893721] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.893885] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894061] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_inbound_addr = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894234] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894398] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894560] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_scheme = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894738] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_timeout_action = abort {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.894909] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_tunnelled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.895091] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_uri = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.895273] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.live_migration_with_native_tls = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.895446] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.max_queues = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.895615] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.895859] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.896041] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.nfs_mount_options = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.897356] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.897557] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.897737] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.897910] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.898143] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.898331] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_pcie_ports = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.898510] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.898683] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.pmem_namespaces = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.898850] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.quobyte_client_cfg = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.899305] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.899519] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.899699] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.899870] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900053] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rbd_secret_uuid = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900226] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rbd_user = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900402] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900581] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900748] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rescue_image_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.900912] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rescue_kernel_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.901116] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rescue_ramdisk_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.901313] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.901483] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.rx_queue_size = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.901656] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.smbfs_mount_options = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.901968] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.902168] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.snapshot_compression = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.902339] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.snapshot_image_format = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.902571] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.902751] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.sparse_logical_volumes = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.902983] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.swtpm_enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.903190] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.swtpm_group = tss {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.903367] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.swtpm_user = tss {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.903541] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.sysinfo_serial = unique {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.903753] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.tb_cache_size = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.903860] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.tx_queue_size = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904050] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.uid_maps = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904250] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.use_virtio_for_bridges = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904432] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.virt_type = kvm {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904605] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.volume_clear = zero {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904771] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.volume_clear_size = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.904941] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.volume_use_multipath = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.905122] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_cache_path = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.905296] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.905465] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.905633] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.905801] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.906128] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.906316] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.vzstorage_mount_user = stack {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.906489] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.906673] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.906869] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.auth_type = password {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907100] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907275] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907448] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907612] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907778] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.907983] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.default_floating_pool = public {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.908181] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.908354] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.extension_sync_interval = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.908525] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.http_retries = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.908690] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.908854] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.909027] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.909213] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.909510] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.909710] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.ovs_bridge = br-int {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.909890] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.physnets = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910080] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.region_name = RegionOne {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910254] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910429] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.service_metadata_proxy = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910594] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910764] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.service_type = network {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.910930] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911111] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911278] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911443] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911637] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911804] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] neutron.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.911984] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] notifications.bdms_in_notifications = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.912230] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] notifications.default_level = INFO {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.912424] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] notifications.notification_format = unversioned {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.912629] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] notifications.notify_on_state_change = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.912825] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913021] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] pci.alias = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913205] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] pci.device_spec = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913376] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] pci.report_in_placement = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913558] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913740] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.auth_type = password {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.913909] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914091] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914259] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914430] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914597] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914763] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.914929] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.default_domain_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.915109] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.default_domain_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.915276] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.domain_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.915440] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.domain_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.915671] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.915869] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916048] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916224] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916391] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916575] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.password = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916742] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.project_domain_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.916947] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.project_domain_name = Default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.917150] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.project_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.917337] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.project_name = service {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.917517] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.region_name = RegionOne {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.917683] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.917851] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918097] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.service_type = placement {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918290] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918457] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918626] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918789] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.system_scope = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.918951] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.919129] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.trust_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.919296] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.user_domain_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.919470] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.user_domain_name = Default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.919790] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.user_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.919980] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.username = nova {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.920193] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.920366] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] placement.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.920557] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.cores = 20 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.920778] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.count_usage_from_placement = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.920997] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.921247] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.injected_file_content_bytes = 10240 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.921493] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.injected_file_path_length = 255 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.921694] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.injected_files = 5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.921874] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.instances = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922063] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.key_pairs = 100 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922248] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.metadata_items = 128 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922420] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.ram = 51200 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922589] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.recheck_quota = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922761] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.server_group_members = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.922934] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] quota.server_groups = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.923130] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.923304] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.923470] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.image_metadata_prefilter = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.923637] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.923804] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.max_attempts = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924025] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.max_placement_results = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924189] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924383] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924554] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924737] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] scheduler.workers = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.924975] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.925187] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.925380] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.925557] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.925727] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.925901] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.926086] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.926296] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.926472] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.host_subset_size = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.926642] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.926809] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.927042] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.927359] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.isolated_hosts = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.927435] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.isolated_images = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.927634] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.927805] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928056] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928265] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.pci_in_placement = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928445] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928614] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928783] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.928953] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.929144] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.929349] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.929553] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.track_instance_changes = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.929888] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.930101] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metrics.required = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.930285] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metrics.weight_multiplier = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.930455] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.930628] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] metrics.weight_setting = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.930975] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.931179] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.931373] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.port_range = 10000:20000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.931550] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.931723] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.931902] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] serial_console.serialproxy_port = 6083 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932094] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932281] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.auth_type = password {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932444] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932604] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932770] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.932936] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933114] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933298] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.send_service_user_token = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933467] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933630] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] service_user.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933802] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.agent_enabled = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.933967] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.934370] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.934581] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.934759] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.html5proxy_port = 6082 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.934965] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.image_compression = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935148] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.jpeg_compression = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935314] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.playback_compression = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935486] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.server_listen = 127.0.0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935659] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935821] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.streaming_mode = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.935981] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] spice.zlib_compression = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.936173] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] upgrade_levels.baseapi = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.936347] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] upgrade_levels.compute = auto {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.936510] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] upgrade_levels.conductor = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.936670] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] upgrade_levels.scheduler = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.936838] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937039] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937216] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937382] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937549] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937713] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.937878] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938057] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938229] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vendordata_dynamic_auth.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938409] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.api_retry_count = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938573] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.ca_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938750] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.938919] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.cluster_name = testcl1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.939109] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.connection_pool_size = 10 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.939280] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.console_delay_seconds = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.939453] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.datastore_regex = ^datastore.* {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.939686] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.939876] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.host_password = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940070] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.host_port = 443 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940239] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.host_username = administrator@vsphere.local {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940416] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.insecure = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940582] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.integration_bridge = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940748] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.maximum_objects = 100 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.940910] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.pbm_default_policy = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941087] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.pbm_enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941251] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.pbm_wsdl_location = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941422] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941584] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.serial_port_proxy_uri = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941742] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.serial_port_service_uri = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.941914] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.task_poll_interval = 0.5 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942103] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.use_linked_clone = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942282] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.vnc_keymap = en-us {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942452] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.vnc_port = 5900 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942617] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vmware.vnc_port_total = 10000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942817] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.auth_schemes = ['none'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.942998] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.943352] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.943548] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.943727] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.novncproxy_port = 6080 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.943916] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.server_listen = 127.0.0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944112] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944290] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.vencrypt_ca_certs = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944452] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.vencrypt_client_cert = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944614] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vnc.vencrypt_client_key = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944805] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.944995] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_deep_image_inspection = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.945189] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.945358] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.945524] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.945692] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.disable_rootwrap = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.945857] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.enable_numa_live_migration = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946077] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946205] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946372] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946537] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.libvirt_disable_apic = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946699] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.946866] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947076] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947257] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947429] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947596] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947758] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.947940] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.948183] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.948370] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.948568] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.948746] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.client_socket_timeout = 900 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.948918] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.default_pool_size = 1000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949110] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.keep_alive = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949311] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.max_header_line = 16384 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949492] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949659] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.ssl_ca_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949821] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.ssl_cert_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.949989] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.ssl_key_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.950225] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.tcp_keepidle = 600 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.950359] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.950527] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] zvm.ca_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.950690] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] zvm.cloud_connector_url = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.951023] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.951213] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] zvm.reachable_timeout = 300 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.951401] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.enforce_new_defaults = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.951810] env[61947]: WARNING oslo_config.cfg [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 513.951999] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.enforce_scope = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.952214] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.policy_default_rule = default {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.952425] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.952605] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.policy_file = policy.yaml {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.952782] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.952948] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953132] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953297] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953463] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953639] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953813] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.953991] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.connection_string = messaging:// {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.954180] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.enabled = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.954388] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.es_doc_type = notification {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.954518] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.es_scroll_size = 10000 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.954686] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.es_scroll_time = 2m {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.954879] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.filter_error_trace = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955078] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.hmac_keys = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955261] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.sentinel_service_name = mymaster {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955434] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.socket_timeout = 0.1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955598] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.trace_requests = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955762] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler.trace_sqlalchemy = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.955945] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler_jaeger.process_tags = {} {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.956127] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler_jaeger.service_name_prefix = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.956294] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] profiler_otlp.service_name_prefix = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.956467] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] remote_debug.host = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.956631] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] remote_debug.port = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.956812] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957018] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957204] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957373] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957547] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957711] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.957873] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958086] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958262] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958441] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958605] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958785] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.958954] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.959150] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.959323] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.959494] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.959661] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.959838] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960009] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960187] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960390] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960521] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960684] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.960854] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961035] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961214] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961381] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961543] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961714] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.961881] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962074] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962256] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962420] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962598] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962769] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.962938] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.963153] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.963329] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_notifications.retry = -1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.963523] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.963706] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.963889] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.auth_section = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964073] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.auth_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964243] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.cafile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964408] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.certfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964578] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.collect_timing = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964736] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.connect_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.964932] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.connect_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965122] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.endpoint_id = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965286] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.endpoint_override = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965449] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.insecure = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965605] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.keyfile = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965763] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.max_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.965918] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.min_version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966088] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.region_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966255] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.retriable_status_codes = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966413] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.service_name = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966573] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.service_type = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966732] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.split_loggers = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.966898] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.status_code_retries = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967104] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.status_code_retry_delay = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967271] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.timeout = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967432] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.valid_interfaces = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967592] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_limit.version = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967760] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_reports.file_event_handler = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.967927] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968127] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] oslo_reports.log_dir = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968316] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968484] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968646] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968821] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.968987] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.969167] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.969342] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.969564] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.969783] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.969971] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.970160] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.970325] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] vif_plug_ovs_privileged.user = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.970552] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.970687] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.970864] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.971087] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.971283] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.971459] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.971631] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.971837] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972075] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972281] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.isolate_vif = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972463] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972636] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972808] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.972981] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.973164] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_vif_ovs.per_port_bridge = False {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.973334] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_brick.lock_path = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.973503] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.973668] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.973843] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.capabilities = [21] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974014] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974193] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.helper_command = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974366] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974533] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974698] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] privsep_osbrick.user = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.974897] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975086] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.group = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975255] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.helper_command = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975425] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975591] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975748] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] nova_sys_admin.user = None {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 513.975884] env[61947]: DEBUG oslo_service.service [None req-fb8c3083-02c1-4d9d-953e-9f848b463034 None None] ******************************************************************************** {{(pid=61947) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 513.976423] env[61947]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 514.479648] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Getting list of instances from cluster (obj){ [ 514.479648] env[61947]: value = "domain-c8" [ 514.479648] env[61947]: _type = "ClusterComputeResource" [ 514.479648] env[61947]: } {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 514.480813] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5f80d0-c6f0-4a08-b7a9-b006e295c178 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.490298] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Got total of 0 instances {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 514.490841] env[61947]: WARNING nova.virt.vmwareapi.driver [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 514.491384] env[61947]: INFO nova.virt.node [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Generated node identity 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 [ 514.491621] env[61947]: INFO nova.virt.node [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Wrote node identity 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 to /opt/stack/data/n-cpu-1/compute_id [ 514.993939] env[61947]: WARNING nova.compute.manager [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Compute nodes ['7263fe00-9b30-4433-9e9f-ec265ed2b8f2'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 516.000371] env[61947]: INFO nova.compute.manager [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 517.005899] env[61947]: WARNING nova.compute.manager [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 517.006257] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.006357] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.006509] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.006662] env[61947]: DEBUG nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 517.007660] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11efa20e-90c6-4c96-84ad-006a6b109a5b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.015782] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3a4a7f-12de-4e4a-8130-d488890a1386 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.028864] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d52dfb-d3f3-4ff8-8208-c8e9a0a820bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.034769] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce52215-dbfc-4c82-8dcb-a8b1962bbce8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.063064] env[61947]: DEBUG nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181502MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 517.063288] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.063524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.565769] env[61947]: WARNING nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] No compute node record for cpu-1:7263fe00-9b30-4433-9e9f-ec265ed2b8f2: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 could not be found. [ 518.069878] env[61947]: INFO nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 [ 519.577794] env[61947]: DEBUG nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 519.578179] env[61947]: DEBUG nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 519.736262] env[61947]: INFO nova.scheduler.client.report [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] [req-355c8ce1-ebcf-4ca8-80f1-910cd9136aff] Created resource provider record via placement API for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 519.753493] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753876b5-a18b-43c4-8384-1cbd4aadc835 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.761481] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbf4cad-edc5-4289-aa5c-f2ebb872a58f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.791542] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6775f2d-e59a-4b53-929f-9acb2da98893 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.798761] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198e42ae-2da2-408e-baf9-e811c50d1174 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.811544] env[61947]: DEBUG nova.compute.provider_tree [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 520.349577] env[61947]: DEBUG nova.scheduler.client.report [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 520.349841] env[61947]: DEBUG nova.compute.provider_tree [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 0 to 1 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 520.349981] env[61947]: DEBUG nova.compute.provider_tree [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 520.401513] env[61947]: DEBUG nova.compute.provider_tree [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 1 to 2 during operation: update_traits {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 520.905945] env[61947]: DEBUG nova.compute.resource_tracker [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 520.906348] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.843s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.906348] env[61947]: DEBUG nova.service [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Creating RPC server for service compute {{(pid=61947) start /opt/stack/nova/nova/service.py:186}} [ 520.920061] env[61947]: DEBUG nova.service [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] Join ServiceGroup membership for this service compute {{(pid=61947) start /opt/stack/nova/nova/service.py:203}} [ 520.920246] env[61947]: DEBUG nova.servicegroup.drivers.db [None req-e6d72958-4615-449f-a638-5e844c2fc67e None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61947) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 545.923860] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 546.427190] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Getting list of instances from cluster (obj){ [ 546.427190] env[61947]: value = "domain-c8" [ 546.427190] env[61947]: _type = "ClusterComputeResource" [ 546.427190] env[61947]: } {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 546.428491] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154fb7c2-924b-41b6-ba0c-6acaaeeafee6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.439104] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Got total of 0 instances {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 546.439353] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 546.439672] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Getting list of instances from cluster (obj){ [ 546.439672] env[61947]: value = "domain-c8" [ 546.439672] env[61947]: _type = "ClusterComputeResource" [ 546.439672] env[61947]: } {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 546.440624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec666488-3421-4f0d-8ebf-fed5528712e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.448438] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Got total of 0 instances {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 560.224478] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquiring lock "051c89d4-d10d-493d-abe9-8daed89fa01a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.224769] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "051c89d4-d10d-493d-abe9-8daed89fa01a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.727940] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.947043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "4a58670f-72ae-48a9-826d-afcff5260047" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.947311] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.187109] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "8686b23b-452a-43a3-881c-bc8446714404" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.187452] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "8686b23b-452a-43a3-881c-bc8446714404" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.282799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.283121] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.285804] env[61947]: INFO nova.compute.claims [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.449647] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.695337] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.977453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.154524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquiring lock "d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.154524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.220203] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.369629] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650c50b2-9889-4b64-aee2-0dd3f6e304d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.379349] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8e1487-2634-4dee-a7b3-f2e3a3f365c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.419085] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea79512-f889-4559-8c56-d50524dbbd38 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.427441] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583ca244-9567-4609-9fe9-5008b51738f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.443740] env[61947]: DEBUG nova.compute.provider_tree [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.658267] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.950025] env[61947]: DEBUG nova.scheduler.client.report [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.194624] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.285698] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.285698] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.329749] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.330248] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.373835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "c0743436-bfd5-4b76-b873-6d0f232772fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.374089] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "c0743436-bfd5-4b76-b873-6d0f232772fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.454213] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.454883] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.457766] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.480s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.459314] env[61947]: INFO nova.compute.claims [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 563.792910] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.837018] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.875832] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.964684] env[61947]: DEBUG nova.compute.utils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.966107] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.966345] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 564.319732] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.368643] env[61947]: DEBUG nova.policy [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fe4bc05577f4c9784869ab20845f5ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e637fd4bd719412c8e1fbe6f0c6a1141', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 564.371424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.404467] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.475876] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.624163] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0605ca48-99e2-49ee-af9b-14d17adcaf59 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.638620] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de45140f-1557-4996-ba06-33b89c14c34a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.681474] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad68624-92e9-4076-92a5-631ce076ad4f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.689917] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523da73d-1c9c-41c3-a49f-6b059686c3dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.706356] env[61947]: DEBUG nova.compute.provider_tree [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.209791] env[61947]: DEBUG nova.scheduler.client.report [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.247684] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Successfully created port: bc0d5d56-5032-49e0-9e98-123b43bad11c {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.488786] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.529217] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.529507] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.529634] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.529800] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.529941] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.531000] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.531000] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.531000] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.531433] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.531605] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.531780] env[61947]: DEBUG nova.virt.hardware [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.532853] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5d1fe0-6947-4a99-a146-b8c18f39772b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.547316] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bcc232-8e92-4d84-a413-9ce7cacff88e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.566546] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f50fe78-216b-447b-a865-be5a7d3b69e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.717729] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.718391] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.726402] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.505s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.726402] env[61947]: INFO nova.compute.claims [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.226450] env[61947]: DEBUG nova.compute.utils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 566.227970] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 566.730743] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.920377] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edde3cc4-f7e7-409d-bec4-ca96fb2e47d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.934826] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc09d4f6-871c-4c1a-b5ed-ec4114ea92eb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.975260] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7c86e2-7c7d-48d9-b073-d6e9874cb8a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.986628] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25c1b3c-ca9c-46a9-b79d-42b6ddedd1c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.004465] env[61947]: DEBUG nova.compute.provider_tree [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.506626] env[61947]: DEBUG nova.scheduler.client.report [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.744808] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.772952] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.773150] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.773318] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.773492] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.773997] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.773997] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.773997] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.774450] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.774648] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.774814] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.774986] env[61947]: DEBUG nova.virt.hardware [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.776525] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c94a13-1e52-4b19-b88b-a463a1a0838a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.785188] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bd3874-0eeb-40bb-9d0e-5dc0cfe21e79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.798807] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 567.815683] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 567.817404] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d88ebae8-df7b-49fd-8887-60488a7e1e1f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.831189] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Created folder: OpenStack in parent group-v4. [ 567.831189] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating folder: Project (cddb4f3eb5094bda969730ef28de0b6e). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 567.831189] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c7471ce-98a0-437b-af88-974cd776cff6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.840778] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Created folder: Project (cddb4f3eb5094bda969730ef28de0b6e) in parent group-v264556. [ 567.841127] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating folder: Instances. Parent ref: group-v264557. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 567.841573] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-045fe1b4-6bf3-4f07-83e2-6b1fea7f476e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.851699] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Created folder: Instances in parent group-v264557. [ 567.851791] env[61947]: DEBUG oslo.service.loopingcall [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.852070] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 567.852331] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a67af020-0fb9-44ab-ac8c-fc31f140d49b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.871658] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 567.871658] env[61947]: value = "task-1224116" [ 567.871658] env[61947]: _type = "Task" [ 567.871658] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.884764] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224116, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.012154] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.012991] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.016386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.822s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.018588] env[61947]: INFO nova.compute.claims [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.367095] env[61947]: ERROR nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 568.367095] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.367095] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.367095] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.367095] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.367095] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.367095] env[61947]: ERROR nova.compute.manager raise self.value [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.367095] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.367095] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.367095] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.367569] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.367569] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.367569] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 568.367569] env[61947]: ERROR nova.compute.manager [ 568.367569] env[61947]: Traceback (most recent call last): [ 568.367712] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.367712] env[61947]: listener.cb(fileno) [ 568.367712] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.367712] env[61947]: result = function(*args, **kwargs) [ 568.367712] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.367712] env[61947]: return func(*args, **kwargs) [ 568.367712] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.367712] env[61947]: raise e [ 568.367712] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.367712] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 568.367712] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.367712] env[61947]: created_port_ids = self._update_ports_for_instance( [ 568.367712] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.367712] env[61947]: with excutils.save_and_reraise_exception(): [ 568.367712] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.367712] env[61947]: self.force_reraise() [ 568.367712] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.367712] env[61947]: raise self.value [ 568.367712] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.367712] env[61947]: updated_port = self._update_port( [ 568.367712] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.367712] env[61947]: _ensure_no_port_binding_failure(port) [ 568.367712] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.367712] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.367712] env[61947]: nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 568.367712] env[61947]: Removing descriptor: 15 [ 568.369721] env[61947]: ERROR nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Traceback (most recent call last): [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] yield resources [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.driver.spawn(context, instance, image_meta, [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] vm_ref = self.build_virtual_machine(instance, [ 568.369721] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] for vif in network_info: [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self._sync_wrapper(fn, *args, **kwargs) [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.wait() [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self[:] = self._gt.wait() [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self._exit_event.wait() [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.370069] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] result = hub.switch() [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self.greenlet.switch() [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] result = function(*args, **kwargs) [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return func(*args, **kwargs) [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise e [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] nwinfo = self.network_api.allocate_for_instance( [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] created_port_ids = self._update_ports_for_instance( [ 568.370396] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] with excutils.save_and_reraise_exception(): [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.force_reraise() [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise self.value [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] updated_port = self._update_port( [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] _ensure_no_port_binding_failure(port) [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise exception.PortBindingFailed(port_id=port['id']) [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 568.370719] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] [ 568.371052] env[61947]: INFO nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Terminating instance [ 568.372578] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquiring lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.372743] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquired lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.373077] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.386753] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224116, 'name': CreateVM_Task, 'duration_secs': 0.311834} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.387394] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 568.388414] env[61947]: DEBUG oslo_vmware.service [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b824994e-e6d0-4b3e-8183-81bb5f0725c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.396112] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.396384] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.396942] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 568.397762] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a0ccf30-6364-46f6-8b01-820fe6cb13e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.403340] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 568.403340] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e5ad8c-7471-261d-d59f-0f80eb340590" [ 568.403340] env[61947]: _type = "Task" [ 568.403340] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.412196] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e5ad8c-7471-261d-d59f-0f80eb340590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.525255] env[61947]: DEBUG nova.compute.utils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.529427] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.529840] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.635128] env[61947]: DEBUG nova.policy [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb1ec1cea23942eeaa1566c36db27a1e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24009a57a33f44f0aa4b6f9ef4180072', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 568.922095] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.922443] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 568.922621] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.922776] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.923197] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 568.923468] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-584672bf-10c0-4d66-a1cd-0817afc538bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.945714] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 568.945902] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 568.946745] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341fd7d6-fe2a-4760-bc2b-246cfb22ce94 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.954593] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ac2a69a-1456-42a3-a1aa-e0347e9c795f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.963244] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 568.963244] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525ff041-5607-33dd-ae76-10f1f9cb581a" [ 568.963244] env[61947]: _type = "Task" [ 568.963244] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.963809] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.976579] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525ff041-5607-33dd-ae76-10f1f9cb581a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.995030] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Successfully created port: bf1d6a15-0043-4c17-a88f-fe4253a90126 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.031571] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.163201] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.180094] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08042dc0-e589-4154-bf53-df00c374cad0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.196647] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac5fb1b-8281-4c1f-9024-2a0d9d3701d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.240552] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7b2827-290c-4883-8ebb-4e822ed003c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.253548] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136067ac-15ca-4bf3-8c53-4f2bb8d3b407 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.269108] env[61947]: DEBUG nova.compute.provider_tree [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.394586] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.394906] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.395105] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 569.395224] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 569.450350] env[61947]: DEBUG nova.compute.manager [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Received event network-changed-bc0d5d56-5032-49e0-9e98-123b43bad11c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 569.450350] env[61947]: DEBUG nova.compute.manager [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Refreshing instance network info cache due to event network-changed-bc0d5d56-5032-49e0-9e98-123b43bad11c. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 569.450350] env[61947]: DEBUG oslo_concurrency.lockutils [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] Acquiring lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.481677] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 569.481940] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating directory with path [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 569.482198] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2375bf80-a19b-4e92-b567-71dc1cd8bd39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.510050] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Created directory with path [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 569.510050] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Fetch image to [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 569.510050] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Downloading image file data 24a0d126-58d4-4780-b581-953c5b860675 to [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk on the data store datastore2 {{(pid=61947) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 569.510625] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b49327-4155-4108-b64b-b36d6527402a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.522680] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ab3853-49b9-47c0-b253-fe58f5187122 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.535150] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e552f4a6-0d7a-427a-908e-fc7f8b3ac960 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.588984] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afae0585-d4c6-4e8d-83f6-b5dc8a115bfb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.603757] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "b35cb8d5-c376-4379-8021-a1e06b09c798" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.604068] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.606272] env[61947]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-b14ebcb5-8c35-489d-92dc-96bff7631d11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.640071] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Downloading image file data 24a0d126-58d4-4780-b581-953c5b860675 to the data store datastore2 {{(pid=61947) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 569.668140] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Releasing lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.668140] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.668140] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 569.668140] env[61947]: DEBUG oslo_concurrency.lockutils [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] Acquired lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.668140] env[61947]: DEBUG nova.network.neutron [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Refreshing network info cache for port bc0d5d56-5032-49e0-9e98-123b43bad11c {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.668540] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2409ed4-6ce7-4f34-8164-24ff68e7e790 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.687831] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6c82c0-f521-4b4f-9dc9-40199851620d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.718493] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 051c89d4-d10d-493d-abe9-8daed89fa01a could not be found. [ 569.718650] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 569.719040] env[61947]: INFO nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 569.719296] env[61947]: DEBUG oslo.service.loopingcall [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 569.721692] env[61947]: DEBUG nova.compute.manager [-] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.721809] env[61947]: DEBUG nova.network.neutron [-] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 569.757924] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 569.831124] env[61947]: DEBUG nova.scheduler.client.report [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.843690] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "bde91666-023a-4a56-a026-d63ed320f0ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.843890] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "bde91666-023a-4a56-a026-d63ed320f0ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.881051] env[61947]: DEBUG nova.network.neutron [-] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.906181] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 569.906181] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 569.909161] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 569.909507] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 569.909679] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Didn't find any instances for network info cache update. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 569.911788] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.912594] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.912820] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.913025] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.913319] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.913586] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 569.913770] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 569.914204] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 570.051721] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.089810] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.090100] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.090282] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.090476] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.090620] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.090760] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.090992] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.093610] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.093776] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.093991] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.094217] env[61947]: DEBUG nova.virt.hardware [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.095488] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d096d62f-9069-45e4-bbbb-17f8b4b149c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.107201] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d289be1-6694-4d89-8abb-6e0c9639f18a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.112144] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.219434] env[61947]: DEBUG nova.network.neutron [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.345873] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.350203] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.353342] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.355610] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.036s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.357069] env[61947]: INFO nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.390991] env[61947]: DEBUG nova.network.neutron [-] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.414584] env[61947]: DEBUG nova.network.neutron [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.418992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.478813] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 570.479029] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 570.597546] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Downloaded image file data 24a0d126-58d4-4780-b581-953c5b860675 to vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk on the data store datastore2 {{(pid=61947) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 570.599227] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 570.599517] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Copying Virtual Disk [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk to [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 570.599805] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d30442af-6c33-4270-b6f8-1ee41b6e3939 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.607654] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 570.607654] env[61947]: value = "task-1224117" [ 570.607654] env[61947]: _type = "Task" [ 570.607654] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.617294] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.652785] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.862145] env[61947]: DEBUG nova.compute.utils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.863711] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.863898] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.896142] env[61947]: INFO nova.compute.manager [-] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Took 1.17 seconds to deallocate network for instance. [ 570.900666] env[61947]: DEBUG nova.compute.claims [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 570.901581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.906031] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.922645] env[61947]: DEBUG oslo_concurrency.lockutils [req-512978f7-1d9c-4db5-9b53-90a521b904dc req-5f18734a-8707-42af-82c4-ca85c256d1cd service nova] Releasing lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.121589] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224117, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.193223] env[61947]: DEBUG nova.policy [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3f19c9e88cc4065bfd3ce406f76fa4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f5d0c9b379743738881d7955f2dea10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 571.367884] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.504529] env[61947]: ERROR nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 571.504529] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.504529] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.504529] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.504529] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.504529] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.504529] env[61947]: ERROR nova.compute.manager raise self.value [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.504529] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.504529] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.504529] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.505047] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.505047] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.505047] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 571.505047] env[61947]: ERROR nova.compute.manager [ 571.505047] env[61947]: Traceback (most recent call last): [ 571.505047] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.505047] env[61947]: listener.cb(fileno) [ 571.505047] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.505047] env[61947]: result = function(*args, **kwargs) [ 571.505047] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.505047] env[61947]: return func(*args, **kwargs) [ 571.505047] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.505047] env[61947]: raise e [ 571.505047] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.505047] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 571.505047] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.505047] env[61947]: created_port_ids = self._update_ports_for_instance( [ 571.505047] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.505047] env[61947]: with excutils.save_and_reraise_exception(): [ 571.505047] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.505047] env[61947]: self.force_reraise() [ 571.505047] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.505047] env[61947]: raise self.value [ 571.505047] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.505047] env[61947]: updated_port = self._update_port( [ 571.505047] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.505047] env[61947]: _ensure_no_port_binding_failure(port) [ 571.505047] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.505047] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.505748] env[61947]: nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 571.505748] env[61947]: Removing descriptor: 15 [ 571.505748] env[61947]: ERROR nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] Traceback (most recent call last): [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] yield resources [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.driver.spawn(context, instance, image_meta, [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.505748] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] vm_ref = self.build_virtual_machine(instance, [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] for vif in network_info: [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self._sync_wrapper(fn, *args, **kwargs) [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.wait() [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self[:] = self._gt.wait() [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self._exit_event.wait() [ 571.506064] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] result = hub.switch() [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self.greenlet.switch() [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] result = function(*args, **kwargs) [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return func(*args, **kwargs) [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise e [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] nwinfo = self.network_api.allocate_for_instance( [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.506543] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] created_port_ids = self._update_ports_for_instance( [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] with excutils.save_and_reraise_exception(): [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.force_reraise() [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise self.value [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] updated_port = self._update_port( [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] _ensure_no_port_binding_failure(port) [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.506883] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise exception.PortBindingFailed(port_id=port['id']) [ 571.507262] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 571.507262] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] [ 571.507262] env[61947]: INFO nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Terminating instance [ 571.514502] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.516057] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquired lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.516057] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.554436] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dbda7e-5a01-4888-9f03-33a1b2013943 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.562304] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a9cec4-0d12-4764-be98-cbf8563b0c29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.595101] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290bc51b-2376-4705-b9a7-66911454a3b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.603897] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccbff7ce-b12c-476d-a1e8-7e60f42095f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.623925] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.638947] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224117, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.111402] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.130085] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.134682] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224117, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.381952] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.400037] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.419279] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.419605] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.419763] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.420348] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.422711] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.422711] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.422711] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.422711] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.422711] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.422901] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.422901] env[61947]: DEBUG nova.virt.hardware [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.423237] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba462ee7-e44f-4ed3-a68b-c4c9e0959cc0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.432746] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab281a7-968c-4cba-ab35-37b416cd1098 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.624021] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224117, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.656629} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.625563] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Copied Virtual Disk [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk to [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 572.625748] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleting the datastore file [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 572.626029] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0e26aa5-f338-43d1-8276-6d43ec0d7816 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.635198] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 572.635198] env[61947]: value = "task-1224118" [ 572.635198] env[61947]: _type = "Task" [ 572.635198] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.638876] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.639481] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.644048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.271s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.644048] env[61947]: INFO nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.652250] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.679123] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Successfully created port: 89f60fb6-4df7-430c-b693-85caaaabc703 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.732465] env[61947]: DEBUG nova.compute.manager [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Received event network-vif-deleted-bc0d5d56-5032-49e0-9e98-123b43bad11c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 572.732673] env[61947]: DEBUG nova.compute.manager [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Received event network-changed-bf1d6a15-0043-4c17-a88f-fe4253a90126 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 572.733067] env[61947]: DEBUG nova.compute.manager [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Refreshing instance network info cache due to event network-changed-bf1d6a15-0043-4c17-a88f-fe4253a90126. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 572.733067] env[61947]: DEBUG oslo_concurrency.lockutils [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] Acquiring lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.909292] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Releasing lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.910253] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 572.910253] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 572.910837] env[61947]: DEBUG oslo_concurrency.lockutils [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] Acquired lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.911071] env[61947]: DEBUG nova.network.neutron [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Refreshing network info cache for port bf1d6a15-0043-4c17-a88f-fe4253a90126 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 572.913586] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f7b8eeb-95e8-46f1-9c7c-a158f7ade82c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.929887] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd995d4-9996-4809-a678-00c12e7d1104 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.957441] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8686b23b-452a-43a3-881c-bc8446714404 could not be found. [ 572.958008] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 572.958345] env[61947]: INFO nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Took 0.05 seconds to destroy the instance on the hypervisor. [ 572.958722] env[61947]: DEBUG oslo.service.loopingcall [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.959049] env[61947]: DEBUG nova.compute.manager [-] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.959180] env[61947]: DEBUG nova.network.neutron [-] [instance: 8686b23b-452a-43a3-881c-bc8446714404] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.008243] env[61947]: DEBUG nova.network.neutron [-] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.154495] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.155702] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.021368} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.159900] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.159900] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.162655] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 573.162655] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Moving file from [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb/24a0d126-58d4-4780-b581-953c5b860675 to [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675. {{(pid=61947) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 573.162655] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-b68934ba-7f0d-4bc6-8f74-1c35906204c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.173623] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 573.173623] env[61947]: value = "task-1224119" [ 573.173623] env[61947]: _type = "Task" [ 573.173623] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.188961] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224119, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.322244] env[61947]: DEBUG nova.policy [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9bc1e4b4ad245d89165007d7b3ef8b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1994e700003a49e0bc204b08ddf2970a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 573.462275] env[61947]: DEBUG nova.network.neutron [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.512444] env[61947]: DEBUG nova.network.neutron [-] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.663609] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 573.687057] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224119, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.028233} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.687385] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] File moved {{(pid=61947) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 573.687561] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Cleaning up location [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 573.688222] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleting the datastore file [datastore2] vmware_temp/79d1192e-a870-4e70-80fd-b0bf6c4332bb {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 573.688222] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa042d4c-5522-47f8-981e-ed3e1a3bf626 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.695573] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 573.695573] env[61947]: value = "task-1224120" [ 573.695573] env[61947]: _type = "Task" [ 573.695573] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.704773] env[61947]: DEBUG nova.network.neutron [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.711191] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.837179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05aed9b-4600-404e-ad34-370bb78c49ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.846712] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95feca0a-1d07-499d-a218-89a2c89db6d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.886242] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78aee2e6-9203-4c5b-8c4f-b8cf8ef2c4c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.895566] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a86548-9b2f-4a48-9870-732691c2911c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.915556] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.014915] env[61947]: INFO nova.compute.manager [-] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Took 1.06 seconds to deallocate network for instance. [ 574.021762] env[61947]: DEBUG nova.compute.claims [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 574.022398] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.209309] env[61947]: DEBUG oslo_concurrency.lockutils [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] Releasing lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.209309] env[61947]: DEBUG nova.compute.manager [req-e393fa43-5d42-4fd0-b04e-b1722a024fd2 req-1abfb568-4196-4d5b-bfba-a0ba3ecfe2f1 service nova] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Received event network-vif-deleted-bf1d6a15-0043-4c17-a88f-fe4253a90126 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 574.209309] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224120, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035619} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.209661] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.211395] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b84454f-937a-4ff2-9c5f-cef44d547b7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.216989] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 574.216989] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525c80d9-5d2e-fbe3-a040-a4ca1e5ce791" [ 574.216989] env[61947]: _type = "Task" [ 574.216989] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.228942] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525c80d9-5d2e-fbe3-a040-a4ca1e5ce791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.422656] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.469587] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Successfully created port: e410b280-7506-4bfb-bcb9-e4d6722046e4 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.682057] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.733963] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525c80d9-5d2e-fbe3-a040-a4ca1e5ce791, 'name': SearchDatastore_Task, 'duration_secs': 0.031745} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.734269] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.734694] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4a58670f-72ae-48a9-826d-afcff5260047/4a58670f-72ae-48a9-826d-afcff5260047.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 574.734981] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-958f1b23-3a15-4de7-8fe9-4ec769a1b62d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.751361] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 574.751361] env[61947]: value = "task-1224121" [ 574.751361] env[61947]: _type = "Task" [ 574.751361] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.769672] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.772995] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.773994] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.774122] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.774260] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.774426] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.774739] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.774839] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.775088] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.775188] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.775315] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.775509] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.776488] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf43ba9b-d91f-4d98-b448-4319833a5681 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.785530] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be081c21-1332-406c-a6c5-740be741099a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.928633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.929522] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.933351] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.529s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.938960] env[61947]: INFO nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.278395] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224121, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.358032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquiring lock "4bf2b6e0-2a96-418b-81fd-1104bf510d4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.358315] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "4bf2b6e0-2a96-418b-81fd-1104bf510d4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.445504] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.456047] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.456238] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.692733] env[61947]: DEBUG nova.policy [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9bc1e4b4ad245d89165007d7b3ef8b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1994e700003a49e0bc204b08ddf2970a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 575.765075] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592323} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.765449] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4a58670f-72ae-48a9-826d-afcff5260047/4a58670f-72ae-48a9-826d-afcff5260047.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 575.765671] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 575.765948] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a4fd5a0-4bfb-43b9-8e3d-b2f70534ec4c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.773590] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 575.773590] env[61947]: value = "task-1224122" [ 575.773590] env[61947]: _type = "Task" [ 575.773590] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.781286] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.867181] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.958167] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.152694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6560ee5e-2177-4ffb-865e-7f62198dfd00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.163664] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc99b9e-7c4a-4221-9b11-43e870d8aca5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.198273] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3c4075-d386-4eb9-95d6-69c68a8c723c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.207389] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce82b650-623f-4daa-a2d4-91a51e89db6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.223973] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.285970] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064488} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.286258] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 576.287148] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba3a38e-3d1b-4ef8-8071-4270282dd205 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.310630] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 4a58670f-72ae-48a9-826d-afcff5260047/4a58670f-72ae-48a9-826d-afcff5260047.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 576.311418] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7161d55-1cbe-4c84-9426-6ea1ef3c521b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.335841] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 576.335841] env[61947]: value = "task-1224123" [ 576.335841] env[61947]: _type = "Task" [ 576.335841] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.344141] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.397514] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.730176] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.849375] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224123, 'name': ReconfigVM_Task, 'duration_secs': 0.360316} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.852019] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 4a58670f-72ae-48a9-826d-afcff5260047/4a58670f-72ae-48a9-826d-afcff5260047.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 576.852019] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8aa8b7c3-4672-4a5d-b2d6-53d7cc518948 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.859065] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 576.859065] env[61947]: value = "task-1224124" [ 576.859065] env[61947]: _type = "Task" [ 576.859065] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.868683] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224124, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.969234] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.002951] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.003618] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.004116] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.004696] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.005228] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.005795] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.006224] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.008021] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.008021] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.008021] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.008021] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.009027] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e55133-1885-4082-8503-9f4536c49d7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.018968] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b191055-0039-46ec-9daf-8ef6a68927a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.236146] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.236146] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.238826] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.820s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.240027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.240027] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 577.240027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.587s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.241338] env[61947]: INFO nova.compute.claims [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.245257] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27fd7ff-8af3-42fd-aaf5-dc1bfc85c6b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.256705] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f399f2c-d149-44d6-9596-44e0b14df9e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.280429] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37422214-e923-42b1-a062-9b2ae5625dc0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.290113] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6b8d72-29f9-4970-b9b3-38a50699c71b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.327499] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181483MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 577.327699] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.370435] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224124, 'name': Rename_Task, 'duration_secs': 0.140994} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.370691] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 577.370928] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5291c996-ad70-4c54-bcd6-f7eb11755b1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.378564] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 577.378564] env[61947]: value = "task-1224125" [ 577.378564] env[61947]: _type = "Task" [ 577.378564] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.387595] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.613251] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Successfully created port: bfe8e83f-b111-480b-9d81-755cd573e9c6 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 577.747859] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.750655] env[61947]: ERROR nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 577.750655] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.750655] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.750655] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.750655] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.750655] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.750655] env[61947]: ERROR nova.compute.manager raise self.value [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.750655] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.750655] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.750655] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.751148] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.751148] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.751148] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 577.751148] env[61947]: ERROR nova.compute.manager [ 577.751148] env[61947]: Traceback (most recent call last): [ 577.751148] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.751148] env[61947]: listener.cb(fileno) [ 577.751148] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.751148] env[61947]: result = function(*args, **kwargs) [ 577.751148] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.751148] env[61947]: return func(*args, **kwargs) [ 577.751148] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.751148] env[61947]: raise e [ 577.751148] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.751148] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 577.751148] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.751148] env[61947]: created_port_ids = self._update_ports_for_instance( [ 577.751148] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.751148] env[61947]: with excutils.save_and_reraise_exception(): [ 577.751148] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.751148] env[61947]: self.force_reraise() [ 577.751148] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.751148] env[61947]: raise self.value [ 577.751148] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.751148] env[61947]: updated_port = self._update_port( [ 577.751148] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.751148] env[61947]: _ensure_no_port_binding_failure(port) [ 577.751148] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.751148] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.751914] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 577.751914] env[61947]: Removing descriptor: 17 [ 577.751914] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.751914] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.758640] env[61947]: ERROR nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Traceback (most recent call last): [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] yield resources [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.driver.spawn(context, instance, image_meta, [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] vm_ref = self.build_virtual_machine(instance, [ 577.758640] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] for vif in network_info: [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self._sync_wrapper(fn, *args, **kwargs) [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.wait() [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self[:] = self._gt.wait() [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self._exit_event.wait() [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.758964] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] result = hub.switch() [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self.greenlet.switch() [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] result = function(*args, **kwargs) [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return func(*args, **kwargs) [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise e [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] nwinfo = self.network_api.allocate_for_instance( [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] created_port_ids = self._update_ports_for_instance( [ 577.759336] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] with excutils.save_and_reraise_exception(): [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.force_reraise() [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise self.value [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] updated_port = self._update_port( [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] _ensure_no_port_binding_failure(port) [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise exception.PortBindingFailed(port_id=port['id']) [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 577.759732] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] [ 577.761069] env[61947]: INFO nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Terminating instance [ 577.765509] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquiring lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.765835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquired lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.765835] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.890173] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224125, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.904791] env[61947]: DEBUG nova.policy [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9bc1e4b4ad245d89165007d7b3ef8b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1994e700003a49e0bc204b08ddf2970a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 578.264340] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.337304] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.390265] env[61947]: DEBUG oslo_vmware.api [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224125, 'name': PowerOnVM_Task, 'duration_secs': 0.994943} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.392926] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 578.393154] env[61947]: INFO nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Took 10.65 seconds to spawn the instance on the hypervisor. [ 578.393443] env[61947]: DEBUG nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.394427] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8e7c87-0a8d-4745-a54e-b970bc533c44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.554628] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a445ac9-4f7d-4729-ac89-283815679a7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.563993] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecde5d5-6db6-4849-be7d-c22bf9142957 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.601586] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d731938d-ac96-4f6a-98b4-024ac0bee2d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.609341] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.616404] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bd6ddc-d19a-42d2-a984-db7d7a1a6cea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.631925] env[61947]: DEBUG nova.compute.provider_tree [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.920771] env[61947]: INFO nova.compute.manager [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Took 16.97 seconds to build instance. [ 579.113856] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Releasing lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.113856] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.113856] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 579.114355] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a36e9dac-8c53-4bcb-b9b3-d7a4db660f9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.132119] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413d6099-97d2-4fc7-94f9-477e6fbbb816 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.150075] env[61947]: DEBUG nova.scheduler.client.report [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.168505] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed could not be found. [ 579.168756] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 579.168945] env[61947]: INFO nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Took 0.06 seconds to destroy the instance on the hypervisor. [ 579.169189] env[61947]: DEBUG oslo.service.loopingcall [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.169478] env[61947]: DEBUG nova.compute.manager [-] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.169583] env[61947]: DEBUG nova.network.neutron [-] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.232064] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Successfully created port: 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.249179] env[61947]: DEBUG nova.network.neutron [-] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.277615] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.308170] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.308170] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.308170] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.308304] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.308304] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.308304] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.308304] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.308304] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.308446] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.308446] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.308446] env[61947]: DEBUG nova.virt.hardware [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.308971] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb7d984-8fb9-4e9e-9ed6-b7cbf000283e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.319421] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8decae05-949d-4bf1-a5ae-533870ad4d06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.428156] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5e5ff642-24eb-4604-b4e5-537114eeda9b tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.480s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.507866] env[61947]: DEBUG nova.compute.manager [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Received event network-changed-89f60fb6-4df7-430c-b693-85caaaabc703 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 579.507866] env[61947]: DEBUG nova.compute.manager [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Refreshing instance network info cache due to event network-changed-89f60fb6-4df7-430c-b693-85caaaabc703. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 579.507866] env[61947]: DEBUG oslo_concurrency.lockutils [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] Acquiring lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.508912] env[61947]: DEBUG oslo_concurrency.lockutils [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] Acquired lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.509450] env[61947]: DEBUG nova.network.neutron [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Refreshing network info cache for port 89f60fb6-4df7-430c-b693-85caaaabc703 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 579.527118] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 579.527118] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.527118] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.527118] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.527118] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.527118] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.527118] env[61947]: ERROR nova.compute.manager raise self.value [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.527118] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.527118] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.527118] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.527558] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.527558] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.527558] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 579.527558] env[61947]: ERROR nova.compute.manager [ 579.527558] env[61947]: Traceback (most recent call last): [ 579.527558] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.527558] env[61947]: listener.cb(fileno) [ 579.527558] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.527558] env[61947]: result = function(*args, **kwargs) [ 579.527558] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.527558] env[61947]: return func(*args, **kwargs) [ 579.527558] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.527558] env[61947]: raise e [ 579.527558] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.527558] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 579.527558] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.527558] env[61947]: created_port_ids = self._update_ports_for_instance( [ 579.527558] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.527558] env[61947]: with excutils.save_and_reraise_exception(): [ 579.527558] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.527558] env[61947]: self.force_reraise() [ 579.527558] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.527558] env[61947]: raise self.value [ 579.527558] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.527558] env[61947]: updated_port = self._update_port( [ 579.527558] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.527558] env[61947]: _ensure_no_port_binding_failure(port) [ 579.527558] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.527558] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.528285] env[61947]: nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 579.528285] env[61947]: Removing descriptor: 15 [ 579.528285] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Traceback (most recent call last): [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] yield resources [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.driver.spawn(context, instance, image_meta, [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.528285] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] vm_ref = self.build_virtual_machine(instance, [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] for vif in network_info: [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self._sync_wrapper(fn, *args, **kwargs) [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.wait() [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self[:] = self._gt.wait() [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self._exit_event.wait() [ 579.528585] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] result = hub.switch() [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self.greenlet.switch() [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] result = function(*args, **kwargs) [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return func(*args, **kwargs) [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise e [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] nwinfo = self.network_api.allocate_for_instance( [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.528929] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] created_port_ids = self._update_ports_for_instance( [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] with excutils.save_and_reraise_exception(): [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.force_reraise() [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise self.value [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] updated_port = self._update_port( [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] _ensure_no_port_binding_failure(port) [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.529279] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise exception.PortBindingFailed(port_id=port['id']) [ 579.529576] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 579.529576] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] [ 579.529576] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Terminating instance [ 579.532660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.532660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.532660] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.656775] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.657837] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.661227] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.760s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.705276] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquiring lock "5060a8b3-2492-4da9-a90d-9a2e5d502092" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.705827] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "5060a8b3-2492-4da9-a90d-9a2e5d502092" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.752692] env[61947]: DEBUG nova.network.neutron [-] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.053282] env[61947]: DEBUG nova.network.neutron [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.099430] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.170687] env[61947]: DEBUG nova.compute.utils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.173833] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 580.211832] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.230488] env[61947]: DEBUG nova.network.neutron [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.255464] env[61947]: INFO nova.compute.manager [-] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Took 1.09 seconds to deallocate network for instance. [ 580.258448] env[61947]: DEBUG nova.compute.claims [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 580.258632] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.385510] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13867a5e-0af2-4238-9a23-8cebec0612ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.396904] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bb9947-20dc-440e-95d0-384b9d2283a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.433568] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.435396] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e80e81-5ccb-42a8-a71a-9662abdcc584 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.446728] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dfeb30-d94d-4a2a-ba73-f02f53165ab9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.461479] env[61947]: DEBUG nova.compute.provider_tree [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.675464] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.732288] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.733977] env[61947]: DEBUG oslo_concurrency.lockutils [req-d014f265-788f-463b-bcec-87ea05be0be1 req-1b869d6d-5e51-465b-9abc-45d545dedbdd service nova] Releasing lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.941977] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.942472] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.942675] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.943024] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-341b25ec-5f49-4927-9a25-6ffe785651fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.959897] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6201a2b9-a058-48d1-8df8-b5886894a992 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.979932] env[61947]: DEBUG nova.scheduler.client.report [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.006013] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7 could not be found. [ 581.006490] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 581.007650] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Took 0.06 seconds to destroy the instance on the hypervisor. [ 581.007650] env[61947]: DEBUG oslo.service.loopingcall [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.008609] env[61947]: DEBUG nova.compute.manager [-] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.008851] env[61947]: DEBUG nova.network.neutron [-] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 581.101278] env[61947]: DEBUG nova.network.neutron [-] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.482442] env[61947]: DEBUG nova.compute.manager [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Received event network-changed-e410b280-7506-4bfb-bcb9-e4d6722046e4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 581.482710] env[61947]: DEBUG nova.compute.manager [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Refreshing instance network info cache due to event network-changed-e410b280-7506-4bfb-bcb9-e4d6722046e4. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 581.482990] env[61947]: DEBUG oslo_concurrency.lockutils [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] Acquiring lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.483264] env[61947]: DEBUG oslo_concurrency.lockutils [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] Acquired lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.483890] env[61947]: DEBUG nova.network.neutron [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Refreshing network info cache for port e410b280-7506-4bfb-bcb9-e4d6722046e4 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 581.490220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.828s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.491122] env[61947]: ERROR nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Traceback (most recent call last): [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.driver.spawn(context, instance, image_meta, [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] vm_ref = self.build_virtual_machine(instance, [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.491122] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] for vif in network_info: [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self._sync_wrapper(fn, *args, **kwargs) [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.wait() [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self[:] = self._gt.wait() [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self._exit_event.wait() [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] result = hub.switch() [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.491437] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return self.greenlet.switch() [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] result = function(*args, **kwargs) [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] return func(*args, **kwargs) [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise e [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] nwinfo = self.network_api.allocate_for_instance( [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] created_port_ids = self._update_ports_for_instance( [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] with excutils.save_and_reraise_exception(): [ 581.491768] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] self.force_reraise() [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise self.value [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] updated_port = self._update_port( [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] _ensure_no_port_binding_failure(port) [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] raise exception.PortBindingFailed(port_id=port['id']) [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] nova.exception.PortBindingFailed: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. [ 581.492103] env[61947]: ERROR nova.compute.manager [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] [ 581.492394] env[61947]: DEBUG nova.compute.utils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 581.498128] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.592s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.500854] env[61947]: INFO nova.compute.claims [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.506254] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquiring lock "5019b061-4e47-4c02-ab3f-fcd3e0cc3007" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.506254] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "5019b061-4e47-4c02-ab3f-fcd3e0cc3007" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.507557] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Build of instance 051c89d4-d10d-493d-abe9-8daed89fa01a was re-scheduled: Binding failed for port bc0d5d56-5032-49e0-9e98-123b43bad11c, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 581.507787] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 581.507938] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquiring lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.508076] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Acquired lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.508232] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.604519] env[61947]: DEBUG nova.network.neutron [-] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.692921] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.724766] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.725043] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.725370] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.726074] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.726074] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.726256] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.726531] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.726844] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.727390] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.727631] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.727710] env[61947]: DEBUG nova.virt.hardware [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.728633] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87c0202-9f69-465b-b2b5-18ec0e6c0dde {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.738274] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b55c11-937a-48b9-a3db-938030467099 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.758822] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 581.764561] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating folder: Project (67477831abd44d088b9818b3f5958425). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 581.764906] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1de3cf49-9172-4d1c-bad5-608d9a1b2f6e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.775919] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Created folder: Project (67477831abd44d088b9818b3f5958425) in parent group-v264556. [ 581.775919] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating folder: Instances. Parent ref: group-v264560. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 581.775919] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0437121f-6bfc-49e9-922f-67f5c00bbb6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.785337] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Created folder: Instances in parent group-v264560. [ 581.786776] env[61947]: DEBUG oslo.service.loopingcall [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.786776] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 581.786776] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac111360-bb70-4c06-a6e5-a59b3e7cbd90 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.804818] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 581.804818] env[61947]: value = "task-1224128" [ 581.804818] env[61947]: _type = "Task" [ 581.804818] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.812392] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224128, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.010040] env[61947]: DEBUG nova.network.neutron [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.047979] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.107080] env[61947]: INFO nova.compute.manager [-] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Took 1.10 seconds to deallocate network for instance. [ 582.111225] env[61947]: DEBUG nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 582.111225] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.235777] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.321414] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224128, 'name': CreateVM_Task, 'duration_secs': 0.314742} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.321414] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 582.321414] env[61947]: DEBUG oslo_vmware.service [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508c3726-d337-4519-9ab1-50b365576490 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.331410] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.331497] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.331889] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 582.332165] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4338db98-35d3-4e45-a371-f9ee90f4460b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.338120] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 582.338120] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5226be24-1c34-0ba8-09a0-5d15f5da6e58" [ 582.338120] env[61947]: _type = "Task" [ 582.338120] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.357049] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5226be24-1c34-0ba8-09a0-5d15f5da6e58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.464301] env[61947]: DEBUG nova.network.neutron [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.579410] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 582.579410] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.579410] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.579410] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.579410] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.579410] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.579410] env[61947]: ERROR nova.compute.manager raise self.value [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.579410] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.579410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.579410] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.580073] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.580073] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.580073] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 582.580073] env[61947]: ERROR nova.compute.manager [ 582.580073] env[61947]: Traceback (most recent call last): [ 582.580073] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.580073] env[61947]: listener.cb(fileno) [ 582.580073] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.580073] env[61947]: result = function(*args, **kwargs) [ 582.580073] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.580073] env[61947]: return func(*args, **kwargs) [ 582.580073] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.580073] env[61947]: raise e [ 582.580073] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.580073] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 582.580073] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.580073] env[61947]: created_port_ids = self._update_ports_for_instance( [ 582.580073] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.580073] env[61947]: with excutils.save_and_reraise_exception(): [ 582.580073] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.580073] env[61947]: self.force_reraise() [ 582.580073] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.580073] env[61947]: raise self.value [ 582.580073] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.580073] env[61947]: updated_port = self._update_port( [ 582.580073] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.580073] env[61947]: _ensure_no_port_binding_failure(port) [ 582.580073] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.580073] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.580853] env[61947]: nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 582.580853] env[61947]: Removing descriptor: 18 [ 582.580853] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Traceback (most recent call last): [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] yield resources [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.driver.spawn(context, instance, image_meta, [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.580853] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] vm_ref = self.build_virtual_machine(instance, [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] for vif in network_info: [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self._sync_wrapper(fn, *args, **kwargs) [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.wait() [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self[:] = self._gt.wait() [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self._exit_event.wait() [ 582.581248] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] result = hub.switch() [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self.greenlet.switch() [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] result = function(*args, **kwargs) [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return func(*args, **kwargs) [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise e [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] nwinfo = self.network_api.allocate_for_instance( [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.581770] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] created_port_ids = self._update_ports_for_instance( [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] with excutils.save_and_reraise_exception(): [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.force_reraise() [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise self.value [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] updated_port = self._update_port( [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] _ensure_no_port_binding_failure(port) [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.582115] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise exception.PortBindingFailed(port_id=port['id']) [ 582.582428] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 582.582428] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] [ 582.582428] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Terminating instance [ 582.586802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.586802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.586802] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.717720] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31931e51-d05e-45b5-992c-9fba4b935175 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.725356] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d022a58e-b94f-48df-8a27-05d0cf1159f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.758253] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Releasing lock "refresh_cache-051c89d4-d10d-493d-abe9-8daed89fa01a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.758729] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 582.758729] env[61947]: DEBUG nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.758906] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.761457] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecb58b6-79e2-4bd1-9bc1-1ac5afa94d0c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.769293] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa750e0c-9134-48ec-98e0-22554dd081f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.782797] env[61947]: DEBUG nova.compute.provider_tree [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.809765] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.834900] env[61947]: DEBUG nova.compute.manager [None req-14cc9b68-1f42-4096-a54c-f44338e86552 tempest-ServerDiagnosticsV248Test-1832989618 tempest-ServerDiagnosticsV248Test-1832989618-project-admin] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.838920] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7bd2cf-c45a-4e3c-abd5-fae2929d0bc8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.853625] env[61947]: INFO nova.compute.manager [None req-14cc9b68-1f42-4096-a54c-f44338e86552 tempest-ServerDiagnosticsV248Test-1832989618 tempest-ServerDiagnosticsV248Test-1832989618-project-admin] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Retrieving diagnostics [ 582.858304] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aaac5ca-561b-4106-9c88-c41ea5e3618f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.862309] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.862920] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 582.863263] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.863471] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.863682] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 582.864786] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f75c777-3b01-46f4-89d6-4fbf624a7ee5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.901942] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 582.901942] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 582.902633] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6abfaf-8af5-40b1-90e4-bd513b7f3077 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.911785] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fda4226-9279-4672-b577-535307276947 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.917412] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 582.917412] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5274bf65-00f8-fc44-2fb1-3b9c5612097b" [ 582.917412] env[61947]: _type = "Task" [ 582.917412] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.926463] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5274bf65-00f8-fc44-2fb1-3b9c5612097b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.969809] env[61947]: DEBUG oslo_concurrency.lockutils [req-17d1bd36-7da6-4321-a33b-4efd4ca11f35 req-8031d286-1335-4a5f-bd89-3de6253a15b2 service nova] Releasing lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.129123] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.287028] env[61947]: DEBUG nova.scheduler.client.report [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.316128] env[61947]: DEBUG nova.network.neutron [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.329029] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquiring lock "35edbd0c-0634-47b9-9386-66233b4dbc7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.329029] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "35edbd0c-0634-47b9-9386-66233b4dbc7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.358341] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 583.358341] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.358341] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.358341] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.358341] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.358341] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.358341] env[61947]: ERROR nova.compute.manager raise self.value [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.358341] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.358341] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.358341] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.358814] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.358814] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.358814] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 583.358814] env[61947]: ERROR nova.compute.manager [ 583.358814] env[61947]: Traceback (most recent call last): [ 583.358814] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.358814] env[61947]: listener.cb(fileno) [ 583.358814] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.358814] env[61947]: result = function(*args, **kwargs) [ 583.358814] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.358814] env[61947]: return func(*args, **kwargs) [ 583.358814] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.358814] env[61947]: raise e [ 583.358814] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.358814] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 583.358814] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.358814] env[61947]: created_port_ids = self._update_ports_for_instance( [ 583.358814] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.358814] env[61947]: with excutils.save_and_reraise_exception(): [ 583.358814] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.358814] env[61947]: self.force_reraise() [ 583.358814] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.358814] env[61947]: raise self.value [ 583.358814] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.358814] env[61947]: updated_port = self._update_port( [ 583.358814] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.358814] env[61947]: _ensure_no_port_binding_failure(port) [ 583.358814] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.358814] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.359696] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 583.359696] env[61947]: Removing descriptor: 17 [ 583.359696] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Traceback (most recent call last): [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] yield resources [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.driver.spawn(context, instance, image_meta, [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.359696] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] vm_ref = self.build_virtual_machine(instance, [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] for vif in network_info: [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self._sync_wrapper(fn, *args, **kwargs) [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.wait() [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self[:] = self._gt.wait() [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self._exit_event.wait() [ 583.360085] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] result = hub.switch() [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self.greenlet.switch() [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] result = function(*args, **kwargs) [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return func(*args, **kwargs) [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise e [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] nwinfo = self.network_api.allocate_for_instance( [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.360458] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] created_port_ids = self._update_ports_for_instance( [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] with excutils.save_and_reraise_exception(): [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.force_reraise() [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise self.value [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] updated_port = self._update_port( [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] _ensure_no_port_binding_failure(port) [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.360874] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise exception.PortBindingFailed(port_id=port['id']) [ 583.361248] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 583.361248] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] [ 583.361248] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Terminating instance [ 583.365015] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.365834] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.365995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.367436] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.428926] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 583.429216] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating directory with path [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 583.429451] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-453bcaf9-dfbf-4c69-b637-57f9e1b82a3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.452854] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Created directory with path [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 583.452854] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Fetch image to [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 583.452854] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Downloading image file data 24a0d126-58d4-4780-b581-953c5b860675 to [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk on the data store datastore1 {{(pid=61947) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 583.452854] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f89defa-7f77-4a1e-a529-5d88ee86c669 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.468301] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b63a634-51e4-41ba-aec6-ef9a7597ec5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.483946] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069af4a2-06d9-482f-95c2-a41fd6c43ebb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.533948] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e56c3c-b437-4ebb-b40c-d75bc2fd0010 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.540588] env[61947]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-bf4a46de-dc6c-4681-88f6-8beaf9fba9d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.563207] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Downloading image file data 24a0d126-58d4-4780-b581-953c5b860675 to the data store datastore1 {{(pid=61947) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 583.647037] env[61947]: DEBUG oslo_vmware.rw_handles [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 583.794604] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.795268] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.798434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.776s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.817363] env[61947]: INFO nova.compute.manager [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] [instance: 051c89d4-d10d-493d-abe9-8daed89fa01a] Took 1.06 seconds to deallocate network for instance. [ 583.868308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.868308] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.868308] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 583.868308] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90cddcdb-7ee0-4986-99c7-0e49272c8287 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.889060] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db40775e-410c-42a6-b7d2-d82ba76f26fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.912873] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.921503] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9 could not be found. [ 583.921503] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 583.921503] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 583.921897] env[61947]: DEBUG oslo.service.loopingcall [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.922817] env[61947]: DEBUG nova.compute.manager [-] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.922932] env[61947]: DEBUG nova.network.neutron [-] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.961702] env[61947]: DEBUG nova.network.neutron [-] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.128461] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.303807] env[61947]: DEBUG nova.compute.utils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.316388] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 584.316388] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 584.317155] env[61947]: DEBUG oslo_vmware.rw_handles [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 584.317785] env[61947]: DEBUG oslo_vmware.rw_handles [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 584.440840] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Downloaded image file data 24a0d126-58d4-4780-b581-953c5b860675 to vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk on the data store datastore1 {{(pid=61947) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 584.443859] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 584.444288] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copying Virtual Disk [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk to [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 584.444446] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e2d5aa6-c8e3-4f4c-b52b-483831de5676 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.464040] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 584.464040] env[61947]: value = "task-1224129" [ 584.464040] env[61947]: _type = "Task" [ 584.464040] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.469229] env[61947]: DEBUG nova.policy [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0337baafe604194a5bc93e99ca8078a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51aa57efe97e453783044286f33d3f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 584.473113] env[61947]: DEBUG nova.network.neutron [-] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.484184] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.607201] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e58002-cb7a-4706-965b-9b20b5ea735c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.615758] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7f0fd9-07c0-4201-86e0-9df34f9785c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.650680] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.651269] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.651394] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 584.651901] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34b45393-55bb-4693-86b0-e3c3cb313ddb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.654787] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17d3abe-3e8e-4f33-89de-5eef0630bb26 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.669739] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153675db-da32-44d1-98dd-28c51ee9a271 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.686303] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da128d5-ed7e-473a-8ca0-24f35860a339 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.697683] env[61947]: DEBUG nova.compute.provider_tree [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.704136] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0743436-bfd5-4b76-b873-6d0f232772fb could not be found. [ 584.704136] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 584.704356] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 584.704463] env[61947]: DEBUG oslo.service.loopingcall [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.704960] env[61947]: DEBUG nova.compute.manager [-] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.705069] env[61947]: DEBUG nova.network.neutron [-] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.764620] env[61947]: DEBUG nova.network.neutron [-] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.817317] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.880368] env[61947]: INFO nova.scheduler.client.report [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Deleted allocations for instance 051c89d4-d10d-493d-abe9-8daed89fa01a [ 584.975550] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224129, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.977559] env[61947]: INFO nova.compute.manager [-] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Took 1.05 seconds to deallocate network for instance. [ 584.981117] env[61947]: DEBUG nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 584.981655] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.206134] env[61947]: DEBUG nova.scheduler.client.report [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.269282] env[61947]: DEBUG nova.network.neutron [-] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.391062] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e7f5762-74fb-4c5e-8b6a-18958a83eaaf tempest-ServersAdminNegativeTestJSON-1580703476 tempest-ServersAdminNegativeTestJSON-1580703476-project-member] Lock "051c89d4-d10d-493d-abe9-8daed89fa01a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.166s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.473040] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667649} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.473040] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copied Virtual Disk [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk to [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 585.473040] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleting the datastore file [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675/tmp-sparse.vmdk {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 585.473270] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40f8ee95-3180-402a-9543-42e6bd270e8b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.484263] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 585.484263] env[61947]: value = "task-1224130" [ 585.484263] env[61947]: _type = "Task" [ 585.484263] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.492997] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224130, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.586191] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Successfully created port: ced5f576-4d66-492e-b68b-064fc7b80395 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.715136] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.715863] env[61947]: ERROR nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] Traceback (most recent call last): [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.driver.spawn(context, instance, image_meta, [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] vm_ref = self.build_virtual_machine(instance, [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.715863] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] for vif in network_info: [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self._sync_wrapper(fn, *args, **kwargs) [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.wait() [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self[:] = self._gt.wait() [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self._exit_event.wait() [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] result = hub.switch() [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.716569] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return self.greenlet.switch() [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] result = function(*args, **kwargs) [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] return func(*args, **kwargs) [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise e [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] nwinfo = self.network_api.allocate_for_instance( [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] created_port_ids = self._update_ports_for_instance( [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] with excutils.save_and_reraise_exception(): [ 585.716892] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] self.force_reraise() [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise self.value [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] updated_port = self._update_port( [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] _ensure_no_port_binding_failure(port) [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] raise exception.PortBindingFailed(port_id=port['id']) [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] nova.exception.PortBindingFailed: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. [ 585.717258] env[61947]: ERROR nova.compute.manager [instance: 8686b23b-452a-43a3-881c-bc8446714404] [ 585.717534] env[61947]: DEBUG nova.compute.utils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.723816] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Build of instance 8686b23b-452a-43a3-881c-bc8446714404 was re-scheduled: Binding failed for port bf1d6a15-0043-4c17-a88f-fe4253a90126, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 585.723816] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 585.723816] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.723816] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquired lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.724158] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.724996] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.328s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.729975] env[61947]: INFO nova.compute.claims [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.772245] env[61947]: INFO nova.compute.manager [-] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Took 1.07 seconds to deallocate network for instance. [ 585.775098] env[61947]: DEBUG nova.compute.claims [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 585.775098] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.831936] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.856856] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.857171] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.857385] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.857590] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.857771] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.857924] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.858378] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.858455] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.858695] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.858807] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.858936] env[61947]: DEBUG nova.virt.hardware [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.859991] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0719db-57ee-4b5f-afa9-5dde3da93a1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.871822] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377dc5df-57e6-4c3f-8f58-67afe2d98be8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.894411] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.995797] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224130, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022126} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.996775] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 585.997366] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Moving file from [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c/24a0d126-58d4-4780-b581-953c5b860675 to [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675. {{(pid=61947) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 585.998162] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-594f9d54-c046-4308-97de-32310603f374 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.009860] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 586.009860] env[61947]: value = "task-1224131" [ 586.009860] env[61947]: _type = "Task" [ 586.009860] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.021640] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224131, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.086226] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquiring lock "4734ece4-05d3-492f-8d1a-2c113bdd557f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.086654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "4734ece4-05d3-492f-8d1a-2c113bdd557f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.244626] env[61947]: DEBUG nova.compute.manager [req-20fc74a4-eb87-47e1-b613-9200e07e1caf req-784a6e56-783e-4dd6-8289-05e0cf735671 service nova] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Received event network-vif-deleted-89f60fb6-4df7-430c-b693-85caaaabc703 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 586.289035] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.418143] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.491868] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.521758] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224131, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026118} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.522389] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] File moved {{(pid=61947) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 586.522748] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Cleaning up location [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 586.523061] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleting the datastore file [datastore1] vmware_temp/8c712d6e-bfa9-4aa2-b13f-f12e7263061c {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 586.523776] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51460725-c6cd-4f2f-af8d-99fdf2a91668 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.530712] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 586.530712] env[61947]: value = "task-1224132" [ 586.530712] env[61947]: _type = "Task" [ 586.530712] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.542156] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.766491] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "041b3d99-1b11-47f5-9339-2fbc25036cb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.766785] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "041b3d99-1b11-47f5-9339-2fbc25036cb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.997866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Releasing lock "refresh_cache-8686b23b-452a-43a3-881c-bc8446714404" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.998156] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 586.998330] env[61947]: DEBUG nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.998515] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.005104] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bd1bd5-8350-4e53-b186-73f29f240a9d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.014585] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def82a53-6862-4d99-8cf8-10690c2e007d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.052031] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.056852] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1168a00d-e5aa-4775-b6ed-61703ec21e5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.060391] env[61947]: DEBUG nova.network.neutron [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.067454] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024916} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.069556] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.071196] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc0557a5-d0f5-460d-90af-dead1fa341ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.074775] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc0df91-108d-42f2-b8c1-ee804678be92 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.089906] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 587.089906] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523f6e93-4fde-1f7c-dcb9-e4fce5ea3c47" [ 587.089906] env[61947]: _type = "Task" [ 587.089906] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.090666] env[61947]: DEBUG nova.compute.provider_tree [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.100339] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523f6e93-4fde-1f7c-dcb9-e4fce5ea3c47, 'name': SearchDatastore_Task, 'duration_secs': 0.008797} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.101126] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.101405] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 587.101659] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6638bf5a-f41a-42d9-8d46-ab9554007604 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.109211] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 587.109211] env[61947]: value = "task-1224133" [ 587.109211] env[61947]: _type = "Task" [ 587.109211] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.116599] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.566277] env[61947]: INFO nova.compute.manager [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 8686b23b-452a-43a3-881c-bc8446714404] Took 0.56 seconds to deallocate network for instance. [ 587.600209] env[61947]: DEBUG nova.scheduler.client.report [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.621850] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459175} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.622178] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 587.622314] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 587.623839] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85238019-ea50-47d5-8b0d-487435f2f0bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.631663] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 587.631663] env[61947]: value = "task-1224134" [ 587.631663] env[61947]: _type = "Task" [ 587.631663] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.645271] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224134, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.846209] env[61947]: DEBUG nova.compute.manager [req-61fdb85d-e307-45b7-ae68-8e7fc7324732 req-3b9b3ec4-6f53-4b94-b5a2-717941e2c80e service nova] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Received event network-vif-deleted-e410b280-7506-4bfb-bcb9-e4d6722046e4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 588.106581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.106581] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.108873] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.781s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.147133] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224134, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065982} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.147591] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 588.148692] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f321676-ab00-47b5-aecf-2549cec4c340 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.176707] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 588.177492] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df26155e-2255-43ed-815c-e6eea9952309 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.201558] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 588.201558] env[61947]: value = "task-1224135" [ 588.201558] env[61947]: _type = "Task" [ 588.201558] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.213097] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224135, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.604694] env[61947]: INFO nova.scheduler.client.report [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Deleted allocations for instance 8686b23b-452a-43a3-881c-bc8446714404 [ 588.617041] env[61947]: DEBUG nova.compute.utils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.621407] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.621652] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.714769] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224135, 'name': ReconfigVM_Task, 'duration_secs': 0.429805} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.716248] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Reconfigured VM instance instance-00000008 to attach disk [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 588.716888] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2276a58c-2cbc-48a9-82b9-cff14d7d0b3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.724324] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 588.724324] env[61947]: value = "task-1224136" [ 588.724324] env[61947]: _type = "Task" [ 588.724324] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.734473] env[61947]: ERROR nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 588.734473] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.734473] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.734473] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.734473] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.734473] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.734473] env[61947]: ERROR nova.compute.manager raise self.value [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.734473] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.734473] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.734473] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.734928] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.734928] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.734928] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 588.734928] env[61947]: ERROR nova.compute.manager [ 588.734928] env[61947]: Traceback (most recent call last): [ 588.734928] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.734928] env[61947]: listener.cb(fileno) [ 588.734928] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.734928] env[61947]: result = function(*args, **kwargs) [ 588.734928] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.734928] env[61947]: return func(*args, **kwargs) [ 588.734928] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.734928] env[61947]: raise e [ 588.734928] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.734928] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 588.734928] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.734928] env[61947]: created_port_ids = self._update_ports_for_instance( [ 588.734928] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.734928] env[61947]: with excutils.save_and_reraise_exception(): [ 588.734928] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.734928] env[61947]: self.force_reraise() [ 588.734928] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.734928] env[61947]: raise self.value [ 588.734928] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.734928] env[61947]: updated_port = self._update_port( [ 588.734928] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.734928] env[61947]: _ensure_no_port_binding_failure(port) [ 588.734928] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.734928] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.736298] env[61947]: nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 588.736298] env[61947]: Removing descriptor: 18 [ 588.741300] env[61947]: ERROR nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Traceback (most recent call last): [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] yield resources [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.driver.spawn(context, instance, image_meta, [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] vm_ref = self.build_virtual_machine(instance, [ 588.741300] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] for vif in network_info: [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self._sync_wrapper(fn, *args, **kwargs) [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.wait() [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self[:] = self._gt.wait() [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self._exit_event.wait() [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.741722] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] result = hub.switch() [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self.greenlet.switch() [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] result = function(*args, **kwargs) [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return func(*args, **kwargs) [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise e [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] nwinfo = self.network_api.allocate_for_instance( [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] created_port_ids = self._update_ports_for_instance( [ 588.742155] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] with excutils.save_and_reraise_exception(): [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.force_reraise() [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise self.value [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] updated_port = self._update_port( [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] _ensure_no_port_binding_failure(port) [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise exception.PortBindingFailed(port_id=port['id']) [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 588.742518] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] [ 588.742881] env[61947]: INFO nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Terminating instance [ 588.742881] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224136, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.742881] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.742881] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.742881] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.745249] env[61947]: DEBUG nova.policy [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cea18e71f5f7494e847c56e444650c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e804f5f7c264c129bbc680eee86c6ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 589.121835] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.125081] env[61947]: DEBUG oslo_concurrency.lockutils [None req-704e57d4-968d-4e73-9eff-261143d329e7 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "8686b23b-452a-43a3-881c-bc8446714404" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.937s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.162853] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4a58670f-72ae-48a9-826d-afcff5260047 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163024] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163158] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163294] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163416] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance c0743436-bfd5-4b76-b873-6d0f232772fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163526] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b35cb8d5-c376-4379-8021-a1e06b09c798 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163639] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance bde91666-023a-4a56-a026-d63ed320f0ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.163751] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4bf2b6e0-2a96-418b-81fd-1104bf510d4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 589.239234] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224136, 'name': Rename_Task, 'duration_secs': 0.227741} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.239619] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 589.239977] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78c4170d-53dd-4e11-ac89-115bd39c2f39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.249384] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 589.249384] env[61947]: value = "task-1224137" [ 589.249384] env[61947]: _type = "Task" [ 589.249384] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.258329] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224137, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.289154] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.409926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "77f108ae-487b-4940-bb81-b1443ccb6e55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.409926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "77f108ae-487b-4940-bb81-b1443ccb6e55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.418262] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquiring lock "e384076d-7699-450c-839f-c36a8abff553" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.418760] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "e384076d-7699-450c-839f-c36a8abff553" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.528523] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.618278] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Successfully created port: 2bd6502f-5ecb-4dde-8d16-e385e6c01342 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.633812] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.671493] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 5060a8b3-2492-4da9-a90d-9a2e5d502092 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.761302] env[61947]: DEBUG oslo_vmware.api [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224137, 'name': PowerOnVM_Task, 'duration_secs': 0.467944} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.761613] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 589.761801] env[61947]: INFO nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Took 8.07 seconds to spawn the instance on the hypervisor. [ 589.761985] env[61947]: DEBUG nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 589.762799] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af05183-ff54-4660-9073-f054f8e30581 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.035580] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.035580] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.035580] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 590.035580] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0444af7-5d13-4d69-a82f-daf1d6e13aa1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.045044] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059ef507-ad5e-494c-8af7-0c3e91a342a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.073033] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bde91666-023a-4a56-a026-d63ed320f0ca could not be found. [ 590.073382] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 590.073634] env[61947]: INFO nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.073884] env[61947]: DEBUG oslo.service.loopingcall [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.074485] env[61947]: DEBUG nova.compute.manager [-] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.074590] env[61947]: DEBUG nova.network.neutron [-] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.100869] env[61947]: DEBUG nova.network.neutron [-] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.132214] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.167718] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.167718] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.168131] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.168131] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.168131] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.168905] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.168905] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.168905] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.169086] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.169367] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.169512] env[61947]: DEBUG nova.virt.hardware [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.170379] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2eac481-6e43-4e39-b9fb-87ae672eceee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.173780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.174462] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 5019b061-4e47-4c02-ab3f-fcd3e0cc3007 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.185812] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ad6d5a-00f8-40e5-a0a4-df9c4f6b390f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.285047] env[61947]: INFO nova.compute.manager [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Took 19.66 seconds to build instance. [ 590.425790] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "5c399af8-6d17-4a74-ae32-dfa54b03a077" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.426199] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "5c399af8-6d17-4a74-ae32-dfa54b03a077" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.604271] env[61947]: DEBUG nova.network.neutron [-] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.680524] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 35edbd0c-0634-47b9-9386-66233b4dbc7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.741505] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Received event network-changed-bfe8e83f-b111-480b-9d81-755cd573e9c6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.742415] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Refreshing instance network info cache due to event network-changed-bfe8e83f-b111-480b-9d81-755cd573e9c6. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 590.742415] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquiring lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.742846] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquired lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.743123] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Refreshing network info cache for port bfe8e83f-b111-480b-9d81-755cd573e9c6 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 590.787573] env[61947]: DEBUG oslo_concurrency.lockutils [None req-974312fd-7b67-40c5-8acd-7225f2d6f073 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.183s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.107714] env[61947]: INFO nova.compute.manager [-] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Took 1.03 seconds to deallocate network for instance. [ 591.113233] env[61947]: DEBUG nova.compute.claims [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 591.113543] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.183670] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4734ece4-05d3-492f-8d1a-2c113bdd557f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.289891] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.330735] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.352680] env[61947]: ERROR nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 591.352680] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.352680] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.352680] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.352680] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.352680] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.352680] env[61947]: ERROR nova.compute.manager raise self.value [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.352680] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.352680] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.352680] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.353227] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.353227] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.353227] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 591.353227] env[61947]: ERROR nova.compute.manager [ 591.353227] env[61947]: Traceback (most recent call last): [ 591.353227] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.353227] env[61947]: listener.cb(fileno) [ 591.353227] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.353227] env[61947]: result = function(*args, **kwargs) [ 591.353227] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.353227] env[61947]: return func(*args, **kwargs) [ 591.353227] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.353227] env[61947]: raise e [ 591.353227] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.353227] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 591.353227] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.353227] env[61947]: created_port_ids = self._update_ports_for_instance( [ 591.353227] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.353227] env[61947]: with excutils.save_and_reraise_exception(): [ 591.353227] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.353227] env[61947]: self.force_reraise() [ 591.353227] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.353227] env[61947]: raise self.value [ 591.353227] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.353227] env[61947]: updated_port = self._update_port( [ 591.353227] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.353227] env[61947]: _ensure_no_port_binding_failure(port) [ 591.353227] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.353227] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.354024] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 591.354024] env[61947]: Removing descriptor: 17 [ 591.354024] env[61947]: ERROR nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Traceback (most recent call last): [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] yield resources [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.driver.spawn(context, instance, image_meta, [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.354024] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] vm_ref = self.build_virtual_machine(instance, [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] for vif in network_info: [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self._sync_wrapper(fn, *args, **kwargs) [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.wait() [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self[:] = self._gt.wait() [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self._exit_event.wait() [ 591.355141] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] result = hub.switch() [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self.greenlet.switch() [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] result = function(*args, **kwargs) [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return func(*args, **kwargs) [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise e [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] nwinfo = self.network_api.allocate_for_instance( [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.355512] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] created_port_ids = self._update_ports_for_instance( [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] with excutils.save_and_reraise_exception(): [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.force_reraise() [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise self.value [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] updated_port = self._update_port( [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] _ensure_no_port_binding_failure(port) [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.355876] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise exception.PortBindingFailed(port_id=port['id']) [ 591.356675] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 591.356675] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] [ 591.356675] env[61947]: INFO nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Terminating instance [ 591.360039] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquiring lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.360039] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquired lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.360039] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.624931] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.687875] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 041b3d99-1b11-47f5-9339-2fbc25036cb7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.746671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "2b7e1437-75a8-43e0-adba-ba470bed6581" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.746897] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2b7e1437-75a8-43e0-adba-ba470bed6581" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.826424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.913598] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.129335] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Releasing lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.129651] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Received event network-vif-deleted-bfe8e83f-b111-480b-9d81-755cd573e9c6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.129894] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Received event network-changed-6e6f9e12-cecf-4f3f-8d6e-099b3819a35e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.130135] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Refreshing instance network info cache due to event network-changed-6e6f9e12-cecf-4f3f-8d6e-099b3819a35e. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 592.130422] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquiring lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.130633] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquired lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.130852] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Refreshing network info cache for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.194444] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance e384076d-7699-450c-839f-c36a8abff553 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.228059] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.698492] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 77f108ae-487b-4940-bb81-b1443ccb6e55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.698492] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 592.698492] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 592.706781] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.728806] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Releasing lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.729659] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.729870] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 592.730859] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16420782-66b9-41ae-b4ee-690f70c077e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.741494] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be24ca25-4210-48ee-ac89-20fb72021d3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.777974] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4bf2b6e0-2a96-418b-81fd-1104bf510d4b could not be found. [ 592.778229] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.778407] env[61947]: INFO nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.778649] env[61947]: DEBUG oslo.service.loopingcall [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.778871] env[61947]: DEBUG nova.compute.manager [-] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.779152] env[61947]: DEBUG nova.network.neutron [-] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.827035] env[61947]: DEBUG nova.network.neutron [-] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.977357] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.051029] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5576c8f2-82c2-4605-b7f1-183eba7212bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.063500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac3f172-53f7-456d-a4a3-e07dcbc631a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.102056] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195f4b43-fb22-4b27-8c2c-8b86899e976a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.110073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566199ec-ff4d-4c70-a391-b669fec7335c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.124391] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.328611] env[61947]: DEBUG nova.network.neutron [-] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.480487] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Releasing lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.481015] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Received event network-vif-deleted-6e6f9e12-cecf-4f3f-8d6e-099b3819a35e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.481154] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Received event network-changed-ced5f576-4d66-492e-b68b-064fc7b80395 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.481315] env[61947]: DEBUG nova.compute.manager [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Refreshing instance network info cache due to event network-changed-ced5f576-4d66-492e-b68b-064fc7b80395. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 593.481980] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquiring lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.481980] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Acquired lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.481980] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Refreshing network info cache for port ced5f576-4d66-492e-b68b-064fc7b80395 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.630487] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.642592] env[61947]: DEBUG nova.compute.manager [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Received event network-vif-deleted-ced5f576-4d66-492e-b68b-064fc7b80395 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.643049] env[61947]: DEBUG nova.compute.manager [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Received event network-changed-2bd6502f-5ecb-4dde-8d16-e385e6c01342 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.643238] env[61947]: DEBUG nova.compute.manager [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Refreshing instance network info cache due to event network-changed-2bd6502f-5ecb-4dde-8d16-e385e6c01342. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 593.643438] env[61947]: DEBUG oslo_concurrency.lockutils [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] Acquiring lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.644147] env[61947]: DEBUG oslo_concurrency.lockutils [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] Acquired lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.644653] env[61947]: DEBUG nova.network.neutron [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Refreshing network info cache for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.832837] env[61947]: INFO nova.compute.manager [-] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Took 1.05 seconds to deallocate network for instance. [ 593.835722] env[61947]: DEBUG nova.compute.claims [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 593.835900] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.901663] env[61947]: INFO nova.compute.manager [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Rebuilding instance [ 593.967111] env[61947]: DEBUG nova.compute.manager [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 593.968944] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b37ce3-e828-432b-8fe5-f9e870dbb5af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.135872] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 594.136165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.028s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.137433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.878s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.213946] env[61947]: DEBUG nova.network.neutron [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.217589] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.485521] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 594.486582] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be2f64f7-2cc4-4e6d-a58b-42299c7f1a96 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.495386] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 594.495386] env[61947]: value = "task-1224138" [ 594.495386] env[61947]: _type = "Task" [ 594.495386] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.507432] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.540306] env[61947]: DEBUG nova.network.neutron [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.546079] env[61947]: DEBUG nova.network.neutron [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.746252] env[61947]: DEBUG nova.compute.manager [None req-e21f14fd-99ab-4313-8ed0-f3c3432f3074 tempest-ServerDiagnosticsV248Test-1832989618 tempest-ServerDiagnosticsV248Test-1832989618-project-admin] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 594.747869] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a1279a-a31f-48d7-bf57-d902691030ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.761665] env[61947]: INFO nova.compute.manager [None req-e21f14fd-99ab-4313-8ed0-f3c3432f3074 tempest-ServerDiagnosticsV248Test-1832989618 tempest-ServerDiagnosticsV248Test-1832989618-project-admin] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Retrieving diagnostics [ 594.762136] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339d939d-22cc-493d-89e2-fe4b3e76740a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.010776] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224138, 'name': PowerOffVM_Task, 'duration_secs': 0.134668} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.011345] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 595.011833] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.012679] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df41a65-7801-4594-a88e-9d3e3ebabc48 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.015911] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9499e4-dc60-4f9f-90bf-4ed3243db405 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.025204] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf2347-8486-4f44-8597-72892d2f471b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.028903] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 595.029483] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdf6b6ca-289d-486a-9a7a-ad505f70d1de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.063598] env[61947]: DEBUG oslo_concurrency.lockutils [req-1143f8cc-2c4a-4307-abff-c414de75b87e req-b1d02097-72b0-42e2-bbdc-f2a9e2910d7a service nova] Releasing lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.064352] env[61947]: DEBUG oslo_concurrency.lockutils [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] Releasing lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.065021] env[61947]: DEBUG nova.compute.manager [req-f17ef52c-1d72-4f3a-8939-0f986201b04b req-4dbf08f9-fef0-429b-b2aa-5784c5f0205c service nova] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Received event network-vif-deleted-2bd6502f-5ecb-4dde-8d16-e385e6c01342 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 595.067010] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa1e481-9390-4b69-9b90-139a026bfd8d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.070199] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 595.070199] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 595.070388] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleting the datastore file [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 595.070563] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3684c1f2-f441-474a-a873-e32f39395fb6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.081677] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59176ec7-40bd-4d1c-b797-f50c9c41efa3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.085347] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 595.085347] env[61947]: value = "task-1224140" [ 595.085347] env[61947]: _type = "Task" [ 595.085347] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.097216] env[61947]: DEBUG nova.compute.provider_tree [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.106878] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.274286] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "24c4904f-aa87-4995-813a-8a591b773532" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.274841] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "24c4904f-aa87-4995-813a-8a591b773532" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.597280] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13543} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.597602] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 595.597750] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 595.598084] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.602216] env[61947]: DEBUG nova.scheduler.client.report [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.113060] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.113060] env[61947]: ERROR nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Traceback (most recent call last): [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.driver.spawn(context, instance, image_meta, [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.113060] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] vm_ref = self.build_virtual_machine(instance, [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] for vif in network_info: [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self._sync_wrapper(fn, *args, **kwargs) [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.wait() [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self[:] = self._gt.wait() [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self._exit_event.wait() [ 596.113537] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] result = hub.switch() [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return self.greenlet.switch() [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] result = function(*args, **kwargs) [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] return func(*args, **kwargs) [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise e [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] nwinfo = self.network_api.allocate_for_instance( [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.113890] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] created_port_ids = self._update_ports_for_instance( [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] with excutils.save_and_reraise_exception(): [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] self.force_reraise() [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise self.value [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] updated_port = self._update_port( [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] _ensure_no_port_binding_failure(port) [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.114227] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] raise exception.PortBindingFailed(port_id=port['id']) [ 596.114548] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] nova.exception.PortBindingFailed: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. [ 596.114548] env[61947]: ERROR nova.compute.manager [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] [ 596.114548] env[61947]: DEBUG nova.compute.utils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.116701] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.384s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.119195] env[61947]: INFO nova.compute.claims [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.123327] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Build of instance d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed was re-scheduled: Binding failed for port 89f60fb6-4df7-430c-b693-85caaaabc703, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.123327] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.123327] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquiring lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.123734] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Acquired lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.123734] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.671280] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.672083] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.672280] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.672344] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.672450] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.672621] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.673023] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.673220] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.673399] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.673566] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.673737] env[61947]: DEBUG nova.virt.hardware [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.674874] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f26ff6-6a0b-4cc5-8534-be30f9c63741 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.679179] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.691554] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b905b5ee-3b4d-4aa9-ac9f-20409b3f95b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.711025] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 596.715960] env[61947]: DEBUG oslo.service.loopingcall [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.716253] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 596.716575] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9079c98-ca69-4c05-b124-fee715e0764a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.738245] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 596.738245] env[61947]: value = "task-1224141" [ 596.738245] env[61947]: _type = "Task" [ 596.738245] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.750650] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224141, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.099657] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.256914] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224141, 'name': CreateVM_Task, 'duration_secs': 0.278591} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.258614] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 597.260261] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.260441] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.260803] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 597.264790] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1326340a-eadd-4617-9c0c-a7f29c4cdf9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.269799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquiring lock "efe14457-b8d8-454e-b0a2-fd087a92b02c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.270048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "efe14457-b8d8-454e-b0a2-fd087a92b02c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.275221] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquiring lock "a759b902-a441-404b-9884-f08c5b8e2a71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.275442] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "a759b902-a441-404b-9884-f08c5b8e2a71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.275703] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 597.275703] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5241d231-0084-7db7-a546-b2ba2f775497" [ 597.275703] env[61947]: _type = "Task" [ 597.275703] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.284525] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5241d231-0084-7db7-a546-b2ba2f775497, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.505220] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c115f5d-ab76-43c2-8421-8fa7f4af4ef4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.516356] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79df13c9-68ef-4d27-8e54-9200133cf003 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.556491] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6007dd-cd38-4a80-b03b-8d00ce169e2a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.566509] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8896b63f-f6a2-4770-8c12-04ad8a337102 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.585773] env[61947]: DEBUG nova.compute.provider_tree [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.602191] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Releasing lock "refresh_cache-d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.602419] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 597.602578] env[61947]: DEBUG nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.602736] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.627473] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.791448] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5241d231-0084-7db7-a546-b2ba2f775497, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.793561] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.793561] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 597.793561] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.793561] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.793814] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 597.793814] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52aff93c-bc39-4142-93a3-88a4c563ce3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.804285] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 597.804285] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 597.804781] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24d4dc33-a02d-4fab-aa6e-10e9df3ab134 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.814269] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 597.814269] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f360b5-d35f-a699-d742-1ce7a8d911a6" [ 597.814269] env[61947]: _type = "Task" [ 597.814269] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.822421] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f360b5-d35f-a699-d742-1ce7a8d911a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.963683] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "4a58670f-72ae-48a9-826d-afcff5260047" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.963780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.964018] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "4a58670f-72ae-48a9-826d-afcff5260047-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.964273] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.964493] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.966832] env[61947]: INFO nova.compute.manager [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Terminating instance [ 597.968654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "refresh_cache-4a58670f-72ae-48a9-826d-afcff5260047" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.968881] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquired lock "refresh_cache-4a58670f-72ae-48a9-826d-afcff5260047" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.968971] env[61947]: DEBUG nova.network.neutron [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.089490] env[61947]: DEBUG nova.scheduler.client.report [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.129246] env[61947]: DEBUG nova.network.neutron [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.331805] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f360b5-d35f-a699-d742-1ce7a8d911a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008486} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.332650] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51fffea9-5a16-4707-8dc6-f214bef68959 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.339273] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 598.339273] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52793a21-9121-82c6-70be-420b109dda33" [ 598.339273] env[61947]: _type = "Task" [ 598.339273] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.351079] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52793a21-9121-82c6-70be-420b109dda33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.498451] env[61947]: DEBUG nova.network.neutron [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.599097] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.599097] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.600271] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.490s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.634127] env[61947]: INFO nova.compute.manager [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] [instance: d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed] Took 1.03 seconds to deallocate network for instance. [ 598.666322] env[61947]: DEBUG nova.network.neutron [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.855274] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52793a21-9121-82c6-70be-420b109dda33, 'name': SearchDatastore_Task, 'duration_secs': 0.008286} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.855274] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.855274] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 598.855660] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea9c5451-d668-40e4-9756-32c3e8fcd091 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.863066] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 598.863066] env[61947]: value = "task-1224142" [ 598.863066] env[61947]: _type = "Task" [ 598.863066] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.874071] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquiring lock "0a685616-f0ca-4de1-a78d-ecb1f40c0802" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.874347] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "0a685616-f0ca-4de1-a78d-ecb1f40c0802" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.880267] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.104058] env[61947]: DEBUG nova.compute.utils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.105740] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.105968] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.169202] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Releasing lock "refresh_cache-4a58670f-72ae-48a9-826d-afcff5260047" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.170044] env[61947]: DEBUG nova.compute.manager [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.170835] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.172187] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd98b005-6b58-4fbe-8d97-4db4b0370cc7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.187160] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 599.187160] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-891f452f-db32-49a5-a9f0-72c5ccb4a3d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.194985] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 599.194985] env[61947]: value = "task-1224143" [ 599.194985] env[61947]: _type = "Task" [ 599.194985] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.206920] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.235495] env[61947]: DEBUG nova.policy [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc4cc2d59740437c86e7bfd138fe0bc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fde2b1c18b7480f99380e665331be6c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 599.382870] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224142, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485746} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.387862] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 599.387983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 599.389890] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a1acf70-2efc-4032-a108-407303f3a33e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.400362] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 599.400362] env[61947]: value = "task-1224144" [ 599.400362] env[61947]: _type = "Task" [ 599.400362] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.411035] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.557382] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c2a96e-df11-4231-bf3a-e7a58306bd20 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.567458] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f084cd-d483-418d-9bf9-42d8ea5ddb45 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.611399] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140468d1-73be-4c32-aa65-e0011f5ff5e7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.614503] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.627017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18dda0da-4580-4df5-8be4-c263a98a33fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.642052] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.687308] env[61947]: INFO nova.scheduler.client.report [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Deleted allocations for instance d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed [ 599.705041] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224143, 'name': PowerOffVM_Task, 'duration_secs': 0.141315} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.705388] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 599.705599] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 599.705992] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67a42683-c4f9-41cb-8fb8-a070b2288e4c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.730485] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 599.730485] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 599.730485] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleting the datastore file [datastore2] 4a58670f-72ae-48a9-826d-afcff5260047 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 599.730485] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28e0fe6f-9f78-4b48-8151-9358c1000084 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.736799] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for the task: (returnval){ [ 599.736799] env[61947]: value = "task-1224146" [ 599.736799] env[61947]: _type = "Task" [ 599.736799] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.746193] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224146, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.912849] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.394632} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.914017] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 599.914961] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da1033e-c5d8-465f-a04b-3f391ee6b4d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.939661] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 599.939993] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecbb3505-d5c8-4eb5-a500-5653cb3dd839 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.961376] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Successfully created port: 7f105d0f-7633-43f6-a03d-4033dbb579bf {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.969855] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 599.969855] env[61947]: value = "task-1224147" [ 599.969855] env[61947]: _type = "Task" [ 599.969855] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.978701] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.017100] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquiring lock "280efc93-f913-4e4a-9ef0-fb4174d7d4cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.017308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "280efc93-f913-4e4a-9ef0-fb4174d7d4cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.144996] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.195468] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d191cfde-4d10-46b5-9bf4-2e0f6161261a tempest-ServerDiagnosticsTest-1578008485 tempest-ServerDiagnosticsTest-1578008485-project-member] Lock "d6d0171d-d3dc-4ed6-87ff-3a4fc38a2bed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.041s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.251227] env[61947]: DEBUG oslo_vmware.api [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Task: {'id': task-1224146, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092545} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.252735] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 600.252735] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 600.252735] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 600.252735] env[61947]: INFO nova.compute.manager [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Took 1.08 seconds to destroy the instance on the hypervisor. [ 600.252735] env[61947]: DEBUG oslo.service.loopingcall [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.253105] env[61947]: DEBUG nova.compute.manager [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.253105] env[61947]: DEBUG nova.network.neutron [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.279712] env[61947]: DEBUG nova.network.neutron [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.481792] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224147, 'name': ReconfigVM_Task, 'duration_secs': 0.291775} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.482311] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Reconfigured VM instance instance-00000008 to attach disk [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798/b35cb8d5-c376-4379-8021-a1e06b09c798.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.483139] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0557bba6-496c-4d7e-80fd-b2b68709a6c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.492990] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 600.492990] env[61947]: value = "task-1224148" [ 600.492990] env[61947]: _type = "Task" [ 600.492990] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.504328] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224148, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.629115] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.650262] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.650905] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Traceback (most recent call last): [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.driver.spawn(context, instance, image_meta, [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] vm_ref = self.build_virtual_machine(instance, [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.650905] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] for vif in network_info: [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self._sync_wrapper(fn, *args, **kwargs) [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.wait() [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self[:] = self._gt.wait() [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self._exit_event.wait() [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] result = hub.switch() [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.653335] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return self.greenlet.switch() [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] result = function(*args, **kwargs) [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] return func(*args, **kwargs) [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise e [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] nwinfo = self.network_api.allocate_for_instance( [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] created_port_ids = self._update_ports_for_instance( [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] with excutils.save_and_reraise_exception(): [ 600.654079] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] self.force_reraise() [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise self.value [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] updated_port = self._update_port( [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] _ensure_no_port_binding_failure(port) [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] raise exception.PortBindingFailed(port_id=port['id']) [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] nova.exception.PortBindingFailed: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. [ 600.654506] env[61947]: ERROR nova.compute.manager [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] [ 600.654838] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.654838] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.672s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.656610] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Build of instance 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7 was re-scheduled: Binding failed for port e410b280-7506-4bfb-bcb9-e4d6722046e4, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.656610] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.657107] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.657107] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.657190] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.660242] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.660543] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.660600] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.660804] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.660924] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.661085] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.661305] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.661457] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.661616] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.661777] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.661944] env[61947]: DEBUG nova.virt.hardware [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.662805] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae486e7-adf8-4121-9b2e-602a9960b241 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.680654] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895697ae-b071-415e-b469-977e465a6ad1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.701344] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 600.705536] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.782737] env[61947]: DEBUG nova.network.neutron [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.982012] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.005504] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224148, 'name': Rename_Task, 'duration_secs': 0.169555} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.005504] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 601.005504] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50673ffc-35c9-46da-a8b5-b507f0390b27 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.011591] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 601.011591] env[61947]: value = "task-1224149" [ 601.011591] env[61947]: _type = "Task" [ 601.011591] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.022726] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.232565] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.284733] env[61947]: INFO nova.compute.manager [-] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Took 1.03 seconds to deallocate network for instance. [ 601.485769] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.486009] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.486208] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.486376] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.518737] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee22e64-40f3-47b2-b2dd-9b0b33b41293 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.528805] env[61947]: DEBUG oslo_vmware.api [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224149, 'name': PowerOnVM_Task, 'duration_secs': 0.422762} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.529488] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.533253] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 601.533253] env[61947]: DEBUG nova.compute.manager [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 601.533980] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957c0c57-e328-467d-b297-e0dee8276007 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.537475] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b8e34d-5ee5-48a9-9afb-8f3ad890f1fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.576304] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de2e960-e650-4301-a6ec-b8b8a3ded080 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.584644] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586802c5-25c0-4cf0-ad93-c9ef2b7b2f62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.601439] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.794937] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.936257] env[61947]: ERROR nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 601.936257] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.936257] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.936257] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.936257] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.936257] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.936257] env[61947]: ERROR nova.compute.manager raise self.value [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.936257] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.936257] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.936257] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.936850] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.936850] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.936850] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 601.936850] env[61947]: ERROR nova.compute.manager [ 601.936850] env[61947]: Traceback (most recent call last): [ 601.936850] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.936850] env[61947]: listener.cb(fileno) [ 601.936850] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.936850] env[61947]: result = function(*args, **kwargs) [ 601.936850] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.936850] env[61947]: return func(*args, **kwargs) [ 601.936850] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.936850] env[61947]: raise e [ 601.936850] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.936850] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 601.936850] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.936850] env[61947]: created_port_ids = self._update_ports_for_instance( [ 601.936850] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.936850] env[61947]: with excutils.save_and_reraise_exception(): [ 601.936850] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.936850] env[61947]: self.force_reraise() [ 601.936850] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.936850] env[61947]: raise self.value [ 601.936850] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.936850] env[61947]: updated_port = self._update_port( [ 601.936850] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.936850] env[61947]: _ensure_no_port_binding_failure(port) [ 601.936850] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.936850] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.938071] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 601.938071] env[61947]: Removing descriptor: 17 [ 601.938071] env[61947]: ERROR nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Traceback (most recent call last): [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] yield resources [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.driver.spawn(context, instance, image_meta, [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.938071] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] vm_ref = self.build_virtual_machine(instance, [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] for vif in network_info: [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self._sync_wrapper(fn, *args, **kwargs) [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.wait() [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self[:] = self._gt.wait() [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self._exit_event.wait() [ 601.938569] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] result = hub.switch() [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self.greenlet.switch() [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] result = function(*args, **kwargs) [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return func(*args, **kwargs) [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise e [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] nwinfo = self.network_api.allocate_for_instance( [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.939076] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] created_port_ids = self._update_ports_for_instance( [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] with excutils.save_and_reraise_exception(): [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.force_reraise() [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise self.value [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] updated_port = self._update_port( [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] _ensure_no_port_binding_failure(port) [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.939504] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise exception.PortBindingFailed(port_id=port['id']) [ 601.940753] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 601.940753] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] [ 601.940753] env[61947]: INFO nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Terminating instance [ 601.940753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquiring lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.940753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquired lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.940753] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.034766] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.062462] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquiring lock "8634de70-f66d-4f54-a13b-226737cb6d4a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.062537] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "8634de70-f66d-4f54-a13b-226737cb6d4a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.086357] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.104739] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.473615] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.540261] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7] Took 1.05 seconds to deallocate network for instance. [ 602.594210] env[61947]: DEBUG nova.compute.manager [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Received event network-changed-7f105d0f-7633-43f6-a03d-4033dbb579bf {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 602.594210] env[61947]: DEBUG nova.compute.manager [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Refreshing instance network info cache due to event network-changed-7f105d0f-7633-43f6-a03d-4033dbb579bf. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 602.594210] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] Acquiring lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.616445] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.616445] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Traceback (most recent call last): [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.driver.spawn(context, instance, image_meta, [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.616445] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] vm_ref = self.build_virtual_machine(instance, [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] for vif in network_info: [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self._sync_wrapper(fn, *args, **kwargs) [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.wait() [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self[:] = self._gt.wait() [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self._exit_event.wait() [ 602.617140] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] result = hub.switch() [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return self.greenlet.switch() [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] result = function(*args, **kwargs) [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] return func(*args, **kwargs) [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise e [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] nwinfo = self.network_api.allocate_for_instance( [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.617575] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] created_port_ids = self._update_ports_for_instance( [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] with excutils.save_and_reraise_exception(): [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] self.force_reraise() [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise self.value [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] updated_port = self._update_port( [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] _ensure_no_port_binding_failure(port) [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.617931] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] raise exception.PortBindingFailed(port_id=port['id']) [ 602.618290] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] nova.exception.PortBindingFailed: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. [ 602.618290] env[61947]: ERROR nova.compute.manager [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] [ 602.618290] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.620146] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Build of instance 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9 was re-scheduled: Binding failed for port bfe8e83f-b111-480b-9d81-755cd573e9c6, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.620588] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.620822] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.620969] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.621238] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.623352] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.849s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.683936] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.738687] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquiring lock "b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.738922] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.168436] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.187737] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Releasing lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.188297] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.188485] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 603.188919] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] Acquired lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.189052] env[61947]: DEBUG nova.network.neutron [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Refreshing network info cache for port 7f105d0f-7633-43f6-a03d-4033dbb579bf {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.190206] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee1c6024-b9e9-42c0-83fe-937f35b5a2fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.205100] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd28aa7-a253-4260-a568-1090f497599b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.230530] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5060a8b3-2492-4da9-a90d-9a2e5d502092 could not be found. [ 603.230760] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 603.231352] env[61947]: INFO nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.231953] env[61947]: DEBUG oslo.service.loopingcall [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.234576] env[61947]: DEBUG nova.compute.manager [-] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.234726] env[61947]: DEBUG nova.network.neutron [-] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.276030] env[61947]: DEBUG nova.network.neutron [-] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.581462] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd4cece-c019-46e5-b857-ea215e772d67 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.590301] env[61947]: INFO nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Deleted allocations for instance 553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7 [ 603.600348] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f63b080-0328-4cf9-96be-ea5b4967a791 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.639408] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8801bb48-435f-4d15-a9ff-68e93a355204 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.648523] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2b2eac-27d2-4558-9df2-16b7c789add6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.666293] env[61947]: DEBUG nova.compute.provider_tree [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.672377] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.729365] env[61947]: DEBUG nova.network.neutron [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.778916] env[61947]: DEBUG nova.network.neutron [-] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.021849] env[61947]: DEBUG nova.network.neutron [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.100695] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "553d4ae0-a2bd-41db-91b2-dfdfa0a8dcd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.815s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.169030] env[61947]: DEBUG nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.176886] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.176886] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 604.176886] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.176886] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.265310] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.281441] env[61947]: INFO nova.compute.manager [-] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Took 1.05 seconds to deallocate network for instance. [ 604.283429] env[61947]: DEBUG nova.compute.claims [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 604.283641] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.345639] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "b35cb8d5-c376-4379-8021-a1e06b09c798" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.345639] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.345639] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "b35cb8d5-c376-4379-8021-a1e06b09c798-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.345639] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.346080] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.351668] env[61947]: INFO nova.compute.manager [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Terminating instance [ 604.353338] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "refresh_cache-b35cb8d5-c376-4379-8021-a1e06b09c798" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.353496] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquired lock "refresh_cache-b35cb8d5-c376-4379-8021-a1e06b09c798" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.353659] env[61947]: DEBUG nova.network.neutron [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.525123] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdded04d-c7b4-41f5-84f4-c34529c68bee req-85af5d38-979e-469c-a157-18e89d5c3002 service nova] Releasing lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.602955] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.674930] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.675618] env[61947]: ERROR nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Traceback (most recent call last): [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.driver.spawn(context, instance, image_meta, [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] vm_ref = self.build_virtual_machine(instance, [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.675618] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] for vif in network_info: [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self._sync_wrapper(fn, *args, **kwargs) [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.wait() [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self[:] = self._gt.wait() [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self._exit_event.wait() [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] result = hub.switch() [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.675998] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return self.greenlet.switch() [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] result = function(*args, **kwargs) [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] return func(*args, **kwargs) [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise e [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] nwinfo = self.network_api.allocate_for_instance( [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] created_port_ids = self._update_ports_for_instance( [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] with excutils.save_and_reraise_exception(): [ 604.676399] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] self.force_reraise() [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise self.value [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] updated_port = self._update_port( [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] _ensure_no_port_binding_failure(port) [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] raise exception.PortBindingFailed(port_id=port['id']) [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] nova.exception.PortBindingFailed: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. [ 604.676786] env[61947]: ERROR nova.compute.manager [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] [ 604.677736] env[61947]: DEBUG nova.compute.utils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.677774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.260s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.679318] env[61947]: INFO nova.compute.claims [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.682425] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Build of instance c0743436-bfd5-4b76-b873-6d0f232772fb was re-scheduled: Binding failed for port 6e6f9e12-cecf-4f3f-8d6e-099b3819a35e, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.682868] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.683102] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquiring lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.683252] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Acquired lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.683411] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.768861] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.900919] env[61947]: DEBUG nova.network.neutron [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.048142] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "6ca2d77f-a62e-4eff-bf15-7855c94bb649" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.048386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "6ca2d77f-a62e-4eff-bf15-7855c94bb649" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.063856] env[61947]: DEBUG nova.network.neutron [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.134939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.238964] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.273693] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9] Took 1.10 seconds to deallocate network for instance. [ 605.341043] env[61947]: DEBUG nova.compute.manager [req-8c4b39bb-4b64-435b-a3df-0df69d4c3a92 req-4771caef-afdc-46f4-8144-f85ee5b96a17 service nova] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Received event network-vif-deleted-7f105d0f-7633-43f6-a03d-4033dbb579bf {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.496333] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.567774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Releasing lock "refresh_cache-b35cb8d5-c376-4379-8021-a1e06b09c798" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.569014] env[61947]: DEBUG nova.compute.manager [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.569014] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 605.569557] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23f8216-a399-493b-acb2-9a334212baa5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.579464] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 605.579625] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c076a66a-b296-495c-a61c-4d9317a2e338 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.587165] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 605.587165] env[61947]: value = "task-1224154" [ 605.587165] env[61947]: _type = "Task" [ 605.587165] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.596021] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.005959] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Releasing lock "refresh_cache-c0743436-bfd5-4b76-b873-6d0f232772fb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.006214] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.007126] env[61947]: DEBUG nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.007126] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.064140] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.118957] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224154, 'name': PowerOffVM_Task, 'duration_secs': 0.134568} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.118957] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 606.118957] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 606.118957] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-458dc1a6-eaa8-44ae-b69e-18dbadb25fa0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.152155] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 606.152961] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 606.152961] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleting the datastore file [datastore1] b35cb8d5-c376-4379-8021-a1e06b09c798 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 606.152961] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2063881e-621a-40f4-b580-5855cfa0913e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.162795] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for the task: (returnval){ [ 606.162795] env[61947]: value = "task-1224156" [ 606.162795] env[61947]: _type = "Task" [ 606.162795] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.171702] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224156, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.236574] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86159ae3-c3f0-49b3-a011-d4a3ff9e0d63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.247734] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed978da-fd41-4117-b71e-7e8b8051d9d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.285144] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9da1c3-2b48-4746-90c8-e5170c0ce57a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.295682] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86b2388-b234-4ae5-8033-959eabd25158 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.310617] env[61947]: DEBUG nova.compute.provider_tree [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.335136] env[61947]: INFO nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Deleted allocations for instance 90d68391-3d3e-4352-86ba-f3aa2ff1f6a9 [ 606.572168] env[61947]: DEBUG nova.network.neutron [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.680070] env[61947]: DEBUG oslo_vmware.api [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Task: {'id': task-1224156, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096203} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.680070] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 606.680070] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 606.680070] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 606.680070] env[61947]: INFO nova.compute.manager [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Took 1.11 seconds to destroy the instance on the hypervisor. [ 606.680504] env[61947]: DEBUG oslo.service.loopingcall [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.680504] env[61947]: DEBUG nova.compute.manager [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.680769] env[61947]: DEBUG nova.network.neutron [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.729022] env[61947]: DEBUG nova.network.neutron [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.815283] env[61947]: DEBUG nova.scheduler.client.report [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.847417] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "90d68391-3d3e-4352-86ba-f3aa2ff1f6a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.517s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.075043] env[61947]: INFO nova.compute.manager [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] [instance: c0743436-bfd5-4b76-b873-6d0f232772fb] Took 1.07 seconds to deallocate network for instance. [ 607.232403] env[61947]: DEBUG nova.network.neutron [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.321505] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.321936] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.326379] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.153s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.330143] env[61947]: INFO nova.compute.claims [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.351890] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.736672] env[61947]: INFO nova.compute.manager [-] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Took 1.06 seconds to deallocate network for instance. [ 607.837855] env[61947]: DEBUG nova.compute.utils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.842947] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.842947] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.896266] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.959674] env[61947]: DEBUG nova.policy [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c62bfeaa5482416a8b423618bf37fc05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '865f2bd192b248baaeb19b0d12fb7c33', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 608.123657] env[61947]: INFO nova.scheduler.client.report [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Deleted allocations for instance c0743436-bfd5-4b76-b873-6d0f232772fb [ 608.248139] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.349683] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.442723] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Successfully created port: 6c4f3643-7770-411e-ae05-1fc6a3d88453 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.636501] env[61947]: DEBUG oslo_concurrency.lockutils [None req-761929f0-9050-4f63-8a39-0cc4eb994f77 tempest-ListServersNegativeTestJSON-1622216044 tempest-ListServersNegativeTestJSON-1622216044-project-member] Lock "c0743436-bfd5-4b76-b873-6d0f232772fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.262s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.682561] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b458d40-3997-444c-bd38-5f623cd66627 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.690914] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3e1fd5-36a0-459c-bbb3-683207485716 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.723022] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dd8e71-4767-41f9-ab20-73238afe3d25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.735508] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb86c7c7-4ee0-4bc9-a40e-4cf81f98135c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.749552] env[61947]: DEBUG nova.compute.provider_tree [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.139748] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.256630] env[61947]: DEBUG nova.scheduler.client.report [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.362150] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.393106] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:45:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='434470709',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1711895530',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.393407] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.393542] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.393695] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.394152] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.394152] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.394453] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.394453] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.394949] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.395182] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.395393] env[61947]: DEBUG nova.virt.hardware [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.396296] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecdf476-82dc-4459-a852-0be03fe65e76 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.404519] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c955193-a2eb-4a10-9303-56277a04919c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.674606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "2eaab038-6c55-439a-8293-da80474d369a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.674991] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "2eaab038-6c55-439a-8293-da80474d369a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.675970] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.768075] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.771024] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.772630] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.659s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.146906] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.146906] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.281360] env[61947]: DEBUG nova.compute.utils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.291151] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.291389] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.562121] env[61947]: DEBUG nova.policy [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd018f3cda4d04dfcbda4d9cb9299d9d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef2f7c648e314cf683a3427b1b9901d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 610.716323] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7617ec-a4ab-474d-b00a-572e351fa8c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.725805] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81e42b8-923b-494a-a346-6bbe7f054f3f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.759504] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de43c2be-8de5-450e-97ec-c42a309a70a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.767452] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20e0d95-ab2e-47b3-a8e7-9cab28835609 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.781975] env[61947]: DEBUG nova.compute.provider_tree [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.791388] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.048834] env[61947]: DEBUG nova.compute.manager [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Received event network-changed-6c4f3643-7770-411e-ae05-1fc6a3d88453 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.049717] env[61947]: DEBUG nova.compute.manager [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Refreshing instance network info cache due to event network-changed-6c4f3643-7770-411e-ae05-1fc6a3d88453. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 611.050048] env[61947]: DEBUG oslo_concurrency.lockutils [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] Acquiring lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.050224] env[61947]: DEBUG oslo_concurrency.lockutils [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] Acquired lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.050363] env[61947]: DEBUG nova.network.neutron [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Refreshing network info cache for port 6c4f3643-7770-411e-ae05-1fc6a3d88453 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.285877] env[61947]: DEBUG nova.scheduler.client.report [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.372234] env[61947]: ERROR nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 611.372234] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.372234] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.372234] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.372234] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.372234] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.372234] env[61947]: ERROR nova.compute.manager raise self.value [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.372234] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.372234] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.372234] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.372751] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.372751] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.372751] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 611.372751] env[61947]: ERROR nova.compute.manager [ 611.375018] env[61947]: Traceback (most recent call last): [ 611.375018] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.375018] env[61947]: listener.cb(fileno) [ 611.375018] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.375018] env[61947]: result = function(*args, **kwargs) [ 611.375018] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.375018] env[61947]: return func(*args, **kwargs) [ 611.375018] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.375018] env[61947]: raise e [ 611.375018] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.375018] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 611.375018] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.375018] env[61947]: created_port_ids = self._update_ports_for_instance( [ 611.375018] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.375018] env[61947]: with excutils.save_and_reraise_exception(): [ 611.375018] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.375018] env[61947]: self.force_reraise() [ 611.375018] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.375018] env[61947]: raise self.value [ 611.375018] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.375018] env[61947]: updated_port = self._update_port( [ 611.375018] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.375018] env[61947]: _ensure_no_port_binding_failure(port) [ 611.375018] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.375018] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.375018] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 611.375018] env[61947]: Removing descriptor: 17 [ 611.375826] env[61947]: ERROR nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Traceback (most recent call last): [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] yield resources [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.driver.spawn(context, instance, image_meta, [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] vm_ref = self.build_virtual_machine(instance, [ 611.375826] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] for vif in network_info: [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self._sync_wrapper(fn, *args, **kwargs) [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.wait() [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self[:] = self._gt.wait() [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self._exit_event.wait() [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.376241] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] result = hub.switch() [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self.greenlet.switch() [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] result = function(*args, **kwargs) [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return func(*args, **kwargs) [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise e [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] nwinfo = self.network_api.allocate_for_instance( [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] created_port_ids = self._update_ports_for_instance( [ 611.376611] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] with excutils.save_and_reraise_exception(): [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.force_reraise() [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise self.value [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] updated_port = self._update_port( [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] _ensure_no_port_binding_failure(port) [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise exception.PortBindingFailed(port_id=port['id']) [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 611.376968] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] [ 611.377427] env[61947]: INFO nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Terminating instance [ 611.380042] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquiring lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.608942] env[61947]: DEBUG nova.network.neutron [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.629797] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.630330] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.794008] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.794824] env[61947]: ERROR nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Traceback (most recent call last): [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.driver.spawn(context, instance, image_meta, [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] vm_ref = self.build_virtual_machine(instance, [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.794824] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] for vif in network_info: [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self._sync_wrapper(fn, *args, **kwargs) [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.wait() [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self[:] = self._gt.wait() [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self._exit_event.wait() [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] result = hub.switch() [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.796942] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return self.greenlet.switch() [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] result = function(*args, **kwargs) [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] return func(*args, **kwargs) [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise e [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] nwinfo = self.network_api.allocate_for_instance( [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] created_port_ids = self._update_ports_for_instance( [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] with excutils.save_and_reraise_exception(): [ 611.797331] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] self.force_reraise() [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise self.value [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] updated_port = self._update_port( [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] _ensure_no_port_binding_failure(port) [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] raise exception.PortBindingFailed(port_id=port['id']) [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] nova.exception.PortBindingFailed: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. [ 611.797725] env[61947]: ERROR nova.compute.manager [instance: bde91666-023a-4a56-a026-d63ed320f0ca] [ 611.798059] env[61947]: DEBUG nova.compute.utils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.798464] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.972s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.800916] env[61947]: INFO nova.compute.claims [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.807740] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Build of instance bde91666-023a-4a56-a026-d63ed320f0ca was re-scheduled: Binding failed for port ced5f576-4d66-492e-b68b-064fc7b80395, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.807740] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.807740] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.807740] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.808045] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.808045] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.839346] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.839346] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.839346] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.839559] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.839592] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.840307] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.840307] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.840307] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.840307] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.840512] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.840672] env[61947]: DEBUG nova.virt.hardware [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.844500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4167625f-7d44-409a-a2d9-a8d44ed6d0df {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.851460] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f577e8-7966-4b4e-add7-3abf8ed270a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.142413] env[61947]: DEBUG nova.network.neutron [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.231503] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Successfully created port: d1eb481c-21d8-4f13-a981-fb6205b17e01 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.386661] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.644966] env[61947]: DEBUG oslo_concurrency.lockutils [req-7d811c2f-d1fb-48b1-bd9b-d24e71e05a13 req-8b29c055-bef4-42f9-aa38-e2229d5c8828 service nova] Releasing lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.645458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquired lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.645738] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.777707] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.208935] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.264826] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2945bd-62d2-4f12-9003-7b7d862cdeab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.278217] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b06458-2736-4174-9363-a656805815ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.291947] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-bde91666-023a-4a56-a026-d63ed320f0ca" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.291947] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.291947] env[61947]: DEBUG nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.291947] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.328789] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26fda3c-167f-4e18-a7a5-052ed3d8d6d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.337194] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7356071-b592-41d4-ab35-e69bee1482b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.354013] env[61947]: DEBUG nova.compute.provider_tree [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.377685] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.565343] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.617108] env[61947]: DEBUG nova.compute.manager [req-98618bcd-bfe1-4da9-8d9c-067995b0cccb req-f373947d-f996-4724-b029-5d66205f8022 service nova] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Received event network-vif-deleted-6c4f3643-7770-411e-ae05-1fc6a3d88453 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 613.854495] env[61947]: DEBUG nova.scheduler.client.report [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.883147] env[61947]: DEBUG nova.network.neutron [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.072671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Releasing lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.073484] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.073799] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 614.077341] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6f30926-fe8b-4be3-a7dc-cac680139e65 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.096359] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea987f7-389d-4cef-8b4f-6f237f69d913 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.129745] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5019b061-4e47-4c02-ab3f-fcd3e0cc3007 could not be found. [ 614.130034] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 614.130180] env[61947]: INFO nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Took 0.06 seconds to destroy the instance on the hypervisor. [ 614.130334] env[61947]: DEBUG oslo.service.loopingcall [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.130578] env[61947]: DEBUG nova.compute.manager [-] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.130781] env[61947]: DEBUG nova.network.neutron [-] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.189288] env[61947]: DEBUG nova.network.neutron [-] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.364320] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.367018] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.370053] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.534s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.388239] env[61947]: INFO nova.compute.manager [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: bde91666-023a-4a56-a026-d63ed320f0ca] Took 1.10 seconds to deallocate network for instance. [ 614.692298] env[61947]: DEBUG nova.network.neutron [-] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.879247] env[61947]: DEBUG nova.compute.utils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.886269] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.886473] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.937318] env[61947]: DEBUG nova.policy [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cc88da7c18409184b3f04904768c3a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd25d95bace314a0c9b505a934cc5fa6a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 615.190557] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquiring lock "d58f3c69-ae7a-48d5-a796-34898c6e704c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.190791] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "d58f3c69-ae7a-48d5-a796-34898c6e704c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.193580] env[61947]: INFO nova.compute.manager [-] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Took 1.06 seconds to deallocate network for instance. [ 615.195674] env[61947]: DEBUG nova.compute.claims [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 615.195674] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.327390] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970d2a10-bb93-4bb5-ad49-80089eb3f7c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.337929] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4f08be-95ef-492c-b079-e4a6b095a122 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.371353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce79ed0f-b2f2-4d46-ac09-d39ace1ea88b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.378032] env[61947]: ERROR nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 615.378032] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.378032] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.378032] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.378032] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.378032] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.378032] env[61947]: ERROR nova.compute.manager raise self.value [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.378032] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.378032] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.378032] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.378579] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.378579] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.378579] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 615.378579] env[61947]: ERROR nova.compute.manager [ 615.378579] env[61947]: Traceback (most recent call last): [ 615.378579] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.378579] env[61947]: listener.cb(fileno) [ 615.378579] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.378579] env[61947]: result = function(*args, **kwargs) [ 615.378579] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.378579] env[61947]: return func(*args, **kwargs) [ 615.378579] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.378579] env[61947]: raise e [ 615.378579] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.378579] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 615.378579] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.378579] env[61947]: created_port_ids = self._update_ports_for_instance( [ 615.378579] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.378579] env[61947]: with excutils.save_and_reraise_exception(): [ 615.378579] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.378579] env[61947]: self.force_reraise() [ 615.378579] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.378579] env[61947]: raise self.value [ 615.378579] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.378579] env[61947]: updated_port = self._update_port( [ 615.378579] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.378579] env[61947]: _ensure_no_port_binding_failure(port) [ 615.378579] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.378579] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.379444] env[61947]: nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 615.379444] env[61947]: Removing descriptor: 18 [ 615.380226] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ea9a0a-ce04-4d93-a490-d6dd41b3b1ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.384537] env[61947]: ERROR nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Traceback (most recent call last): [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] yield resources [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.driver.spawn(context, instance, image_meta, [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] vm_ref = self.build_virtual_machine(instance, [ 615.384537] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] for vif in network_info: [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self._sync_wrapper(fn, *args, **kwargs) [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.wait() [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self[:] = self._gt.wait() [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self._exit_event.wait() [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.384953] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] result = hub.switch() [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self.greenlet.switch() [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] result = function(*args, **kwargs) [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return func(*args, **kwargs) [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise e [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] nwinfo = self.network_api.allocate_for_instance( [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] created_port_ids = self._update_ports_for_instance( [ 615.385388] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] with excutils.save_and_reraise_exception(): [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.force_reraise() [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise self.value [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] updated_port = self._update_port( [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] _ensure_no_port_binding_failure(port) [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise exception.PortBindingFailed(port_id=port['id']) [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 615.386093] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] [ 615.386461] env[61947]: INFO nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Terminating instance [ 615.386893] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.389483] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquiring lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.389723] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquired lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.389952] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.403679] env[61947]: DEBUG nova.compute.provider_tree [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.432667] env[61947]: INFO nova.scheduler.client.report [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocations for instance bde91666-023a-4a56-a026-d63ed320f0ca [ 615.733395] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Successfully created port: 9d5cbd0c-256c-4648-84ea-00bcc757e332 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.913066] env[61947]: DEBUG nova.scheduler.client.report [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.943175] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d681ff3-439a-4fc5-b296-fc57f9601aaf tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "bde91666-023a-4a56-a026-d63ed320f0ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.098s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.961583] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.096175] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.413131] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.418835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.049s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.419949] env[61947]: ERROR nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Traceback (most recent call last): [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.driver.spawn(context, instance, image_meta, [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] vm_ref = self.build_virtual_machine(instance, [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.419949] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] for vif in network_info: [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self._sync_wrapper(fn, *args, **kwargs) [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.wait() [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self[:] = self._gt.wait() [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self._exit_event.wait() [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] result = hub.switch() [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.420725] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return self.greenlet.switch() [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] result = function(*args, **kwargs) [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] return func(*args, **kwargs) [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise e [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] nwinfo = self.network_api.allocate_for_instance( [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] created_port_ids = self._update_ports_for_instance( [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] with excutils.save_and_reraise_exception(): [ 616.421245] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] self.force_reraise() [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise self.value [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] updated_port = self._update_port( [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] _ensure_no_port_binding_failure(port) [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] raise exception.PortBindingFailed(port_id=port['id']) [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] nova.exception.PortBindingFailed: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. [ 616.421608] env[61947]: ERROR nova.compute.manager [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] [ 616.422014] env[61947]: DEBUG nova.compute.utils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.422462] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.190s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.424086] env[61947]: INFO nova.compute.claims [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.427622] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Build of instance 4bf2b6e0-2a96-418b-81fd-1104bf510d4b was re-scheduled: Binding failed for port 2bd6502f-5ecb-4dde-8d16-e385e6c01342, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.428206] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.428381] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquiring lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.428528] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Acquired lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.428747] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.448837] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.455682] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.455987] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.456800] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.457113] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.457356] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.457643] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.458114] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.458525] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.458754] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.459269] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.459269] env[61947]: DEBUG nova.virt.hardware [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.460606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b258b0-1e7e-4939-8882-bd9c91aa01f0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.470086] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46dd91c0-273a-473b-a3fe-e49e929268d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.529492] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquiring lock "947b697d-4120-491d-83ce-21681974c224" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.529746] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "947b697d-4120-491d-83ce-21681974c224" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.577768] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "c8c66631-1fae-4a03-810f-66d64a75b118" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.578560] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "c8c66631-1fae-4a03-810f-66d64a75b118" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.600626] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Releasing lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.601072] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.601700] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.601700] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0039b83a-dd01-414e-a668-48bead8b0d64 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.612547] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478501e7-441c-4254-994f-153ad5de89ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.637937] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.638209] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.638681] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35edbd0c-0634-47b9-9386-66233b4dbc7d could not be found. [ 616.638876] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 616.639063] env[61947]: INFO nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.639302] env[61947]: DEBUG oslo.service.loopingcall [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.639492] env[61947]: DEBUG nova.compute.manager [-] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.639587] env[61947]: DEBUG nova.network.neutron [-] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.668210] env[61947]: DEBUG nova.network.neutron [-] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.752371] env[61947]: DEBUG nova.compute.manager [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Received event network-changed-d1eb481c-21d8-4f13-a981-fb6205b17e01 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.752371] env[61947]: DEBUG nova.compute.manager [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Refreshing instance network info cache due to event network-changed-d1eb481c-21d8-4f13-a981-fb6205b17e01. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 616.752371] env[61947]: DEBUG oslo_concurrency.lockutils [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] Acquiring lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.752371] env[61947]: DEBUG oslo_concurrency.lockutils [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] Acquired lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.752371] env[61947]: DEBUG nova.network.neutron [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Refreshing network info cache for port d1eb481c-21d8-4f13-a981-fb6205b17e01 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.965150] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.975939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.102288] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.169603] env[61947]: DEBUG nova.network.neutron [-] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.283832] env[61947]: DEBUG nova.network.neutron [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.366408] env[61947]: ERROR nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 617.366408] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.366408] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.366408] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.366408] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.366408] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.366408] env[61947]: ERROR nova.compute.manager raise self.value [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.366408] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.366408] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.366408] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.366997] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.366997] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.366997] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 617.366997] env[61947]: ERROR nova.compute.manager [ 617.366997] env[61947]: Traceback (most recent call last): [ 617.366997] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.366997] env[61947]: listener.cb(fileno) [ 617.366997] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.366997] env[61947]: result = function(*args, **kwargs) [ 617.366997] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.366997] env[61947]: return func(*args, **kwargs) [ 617.366997] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.366997] env[61947]: raise e [ 617.366997] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.366997] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 617.366997] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.366997] env[61947]: created_port_ids = self._update_ports_for_instance( [ 617.366997] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.366997] env[61947]: with excutils.save_and_reraise_exception(): [ 617.366997] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.366997] env[61947]: self.force_reraise() [ 617.366997] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.366997] env[61947]: raise self.value [ 617.366997] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.366997] env[61947]: updated_port = self._update_port( [ 617.366997] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.366997] env[61947]: _ensure_no_port_binding_failure(port) [ 617.366997] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.366997] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.368185] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 617.368185] env[61947]: Removing descriptor: 17 [ 617.368185] env[61947]: ERROR nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Traceback (most recent call last): [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] yield resources [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.driver.spawn(context, instance, image_meta, [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.368185] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] vm_ref = self.build_virtual_machine(instance, [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] for vif in network_info: [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self._sync_wrapper(fn, *args, **kwargs) [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.wait() [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self[:] = self._gt.wait() [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self._exit_event.wait() [ 617.368735] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] result = hub.switch() [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self.greenlet.switch() [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] result = function(*args, **kwargs) [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return func(*args, **kwargs) [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise e [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] nwinfo = self.network_api.allocate_for_instance( [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.369210] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] created_port_ids = self._update_ports_for_instance( [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] with excutils.save_and_reraise_exception(): [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.force_reraise() [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise self.value [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] updated_port = self._update_port( [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] _ensure_no_port_binding_failure(port) [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.369629] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise exception.PortBindingFailed(port_id=port['id']) [ 617.369999] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 617.369999] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] [ 617.369999] env[61947]: INFO nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Terminating instance [ 617.370961] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquiring lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.371135] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquired lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.371297] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.381035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "8a4df3f1-67f3-404d-953f-caa6ba1b1641" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.381256] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "8a4df3f1-67f3-404d-953f-caa6ba1b1641" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.384980] env[61947]: DEBUG nova.network.neutron [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.605554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Releasing lock "refresh_cache-4bf2b6e0-2a96-418b-81fd-1104bf510d4b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.607935] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.607935] env[61947]: DEBUG nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.607935] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.630852] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.673395] env[61947]: INFO nova.compute.manager [-] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Took 1.03 seconds to deallocate network for instance. [ 617.675553] env[61947]: DEBUG nova.compute.claims [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 617.675722] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.888032] env[61947]: DEBUG oslo_concurrency.lockutils [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] Releasing lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.888352] env[61947]: DEBUG nova.compute.manager [req-8a23e1b2-0548-4ec4-8100-fd3a217a2f41 req-59a87d9e-dda0-4b52-a48e-220a04dfbcc4 service nova] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Received event network-vif-deleted-d1eb481c-21d8-4f13-a981-fb6205b17e01 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.902536] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcae01b9-eb7e-4c95-ab0b-f0c49c02411c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.910783] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42739c9e-ce00-4fad-8478-385700c67326 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.947764] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.950306] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d11c780-4faa-49d6-ada1-dde680f2f8d4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.957923] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cd829a-6320-4df4-9c27-a4c4458f99ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.976694] env[61947]: DEBUG nova.compute.provider_tree [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.134994] env[61947]: DEBUG nova.network.neutron [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.165803] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.479301] env[61947]: DEBUG nova.scheduler.client.report [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.639144] env[61947]: INFO nova.compute.manager [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] [instance: 4bf2b6e0-2a96-418b-81fd-1104bf510d4b] Took 1.03 seconds to deallocate network for instance. [ 618.668921] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Releasing lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.669313] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.669583] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 618.671051] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b994b5ed-93f1-4a9a-ba06-b06c6cd14bd5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.681665] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29726d5-5d8a-495f-a3ff-c929ee70bc60 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.706666] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4734ece4-05d3-492f-8d1a-2c113bdd557f could not be found. [ 618.706666] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 618.706666] env[61947]: INFO nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.706666] env[61947]: DEBUG oslo.service.loopingcall [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.706666] env[61947]: DEBUG nova.compute.manager [-] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.706666] env[61947]: DEBUG nova.network.neutron [-] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.739023] env[61947]: DEBUG nova.network.neutron [-] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.777086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "8c1101f6-4263-47db-bbd6-2be773ea3380" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.777424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "8c1101f6-4263-47db-bbd6-2be773ea3380" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.887424] env[61947]: DEBUG nova.compute.manager [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Received event network-changed-9d5cbd0c-256c-4648-84ea-00bcc757e332 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.887869] env[61947]: DEBUG nova.compute.manager [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Refreshing instance network info cache due to event network-changed-9d5cbd0c-256c-4648-84ea-00bcc757e332. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 618.889038] env[61947]: DEBUG oslo_concurrency.lockutils [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] Acquiring lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.889158] env[61947]: DEBUG oslo_concurrency.lockutils [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] Acquired lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.889343] env[61947]: DEBUG nova.network.neutron [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Refreshing network info cache for port 9d5cbd0c-256c-4648-84ea-00bcc757e332 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.987036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.987036] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 618.988290] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.193s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.988643] env[61947]: DEBUG nova.objects.instance [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lazy-loading 'resources' on Instance uuid 4a58670f-72ae-48a9-826d-afcff5260047 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 619.241371] env[61947]: DEBUG nova.network.neutron [-] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.420961] env[61947]: DEBUG nova.network.neutron [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.491570] env[61947]: DEBUG nova.compute.utils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.495979] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.496268] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.688583] env[61947]: INFO nova.scheduler.client.report [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Deleted allocations for instance 4bf2b6e0-2a96-418b-81fd-1104bf510d4b [ 619.744566] env[61947]: INFO nova.compute.manager [-] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Took 1.04 seconds to deallocate network for instance. [ 619.747256] env[61947]: DEBUG nova.compute.claims [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 619.747324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.827018] env[61947]: DEBUG nova.policy [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0aee4169c65243af802c3dcc0d27acc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96a19be770ea41a6a339316ae49570ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 619.860918] env[61947]: DEBUG nova.network.neutron [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.993112] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6486909-2eb7-4a85-9e29-1fade2dd4605 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.001264] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.005218] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd41dac-a6f6-4fda-9d66-727b234d500f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.041521] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d942f6e-b78f-49ac-ab6c-3d2c76b1981a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.052606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d70844-7f1c-445d-800c-7528636cad2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.067950] env[61947]: DEBUG nova.compute.provider_tree [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.200704] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8b371ef2-7e82-42eb-8439-11e4957e5768 tempest-VolumesAssistedSnapshotsTest-1294934105 tempest-VolumesAssistedSnapshotsTest-1294934105-project-member] Lock "4bf2b6e0-2a96-418b-81fd-1104bf510d4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.842s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.347280] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Successfully created port: f599e83f-e3a3-42fc-af49-1870ab532e0e {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.373551] env[61947]: DEBUG oslo_concurrency.lockutils [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] Releasing lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.373839] env[61947]: DEBUG nova.compute.manager [req-d880a5e4-c8cb-4b8b-ab32-508ff726a848 req-23f48189-db80-4267-976f-a45237b61283 service nova] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Received event network-vif-deleted-9d5cbd0c-256c-4648-84ea-00bcc757e332 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 620.577195] env[61947]: DEBUG nova.scheduler.client.report [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.703121] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.017874] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.045500] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.045787] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.045881] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.046063] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.046202] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.046338] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.046567] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.046729] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.048412] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.048412] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.048412] env[61947]: DEBUG nova.virt.hardware [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.049873] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1867c17b-2d07-46f6-8eb2-95330674e590 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.058804] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ed85ac-24f0-4e4c-a14a-a2026566d303 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.082774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.094s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.087021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.999s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.087021] env[61947]: DEBUG nova.objects.instance [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61947) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 621.118308] env[61947]: INFO nova.scheduler.client.report [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Deleted allocations for instance 4a58670f-72ae-48a9-826d-afcff5260047 [ 621.225798] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.629143] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c25a1bf-2350-458f-9a88-e34406a9ce2e tempest-ServerDiagnosticsV248Test-1247257033 tempest-ServerDiagnosticsV248Test-1247257033-project-member] Lock "4a58670f-72ae-48a9-826d-afcff5260047" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.665s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.061995] env[61947]: ERROR nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 622.061995] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.061995] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.061995] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.061995] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.061995] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.061995] env[61947]: ERROR nova.compute.manager raise self.value [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.061995] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.061995] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.061995] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.063139] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.063139] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.063139] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 622.063139] env[61947]: ERROR nova.compute.manager [ 622.063139] env[61947]: Traceback (most recent call last): [ 622.063139] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.063139] env[61947]: listener.cb(fileno) [ 622.063139] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.063139] env[61947]: result = function(*args, **kwargs) [ 622.063139] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.063139] env[61947]: return func(*args, **kwargs) [ 622.063139] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.063139] env[61947]: raise e [ 622.063139] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.063139] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 622.063139] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.063139] env[61947]: created_port_ids = self._update_ports_for_instance( [ 622.063139] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.063139] env[61947]: with excutils.save_and_reraise_exception(): [ 622.063139] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.063139] env[61947]: self.force_reraise() [ 622.063139] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.063139] env[61947]: raise self.value [ 622.063139] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.063139] env[61947]: updated_port = self._update_port( [ 622.063139] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.063139] env[61947]: _ensure_no_port_binding_failure(port) [ 622.063139] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.063139] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.064066] env[61947]: nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 622.064066] env[61947]: Removing descriptor: 17 [ 622.064066] env[61947]: ERROR nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Traceback (most recent call last): [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] yield resources [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.driver.spawn(context, instance, image_meta, [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.064066] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] vm_ref = self.build_virtual_machine(instance, [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] for vif in network_info: [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self._sync_wrapper(fn, *args, **kwargs) [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.wait() [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self[:] = self._gt.wait() [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self._exit_event.wait() [ 622.064451] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] result = hub.switch() [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self.greenlet.switch() [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] result = function(*args, **kwargs) [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return func(*args, **kwargs) [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise e [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] nwinfo = self.network_api.allocate_for_instance( [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.064896] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] created_port_ids = self._update_ports_for_instance( [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] with excutils.save_and_reraise_exception(): [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.force_reraise() [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise self.value [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] updated_port = self._update_port( [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] _ensure_no_port_binding_failure(port) [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.065978] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise exception.PortBindingFailed(port_id=port['id']) [ 622.066608] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 622.066608] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] [ 622.066608] env[61947]: INFO nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Terminating instance [ 622.066608] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.066608] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquired lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.066608] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.096955] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d42fc14f-411e-4646-8ef9-c8830f6beea2 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.099939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.814s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.129962] env[61947]: DEBUG nova.compute.manager [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Received event network-changed-f599e83f-e3a3-42fc-af49-1870ab532e0e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 622.130243] env[61947]: DEBUG nova.compute.manager [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Refreshing instance network info cache due to event network-changed-f599e83f-e3a3-42fc-af49-1870ab532e0e. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 622.130369] env[61947]: DEBUG oslo_concurrency.lockutils [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] Acquiring lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.206447] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquiring lock "94d9244e-bf54-442d-8947-71d15d3da551" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.206447] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "94d9244e-bf54-442d-8947-71d15d3da551" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.601139] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.769390] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.811788] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquiring lock "30b75b32-6fde-4a81-af55-11b51f3e087d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.812023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "30b75b32-6fde-4a81-af55-11b51f3e087d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.065239] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcaa2f7-925a-4907-af4c-dd62df0e4a73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.077872] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb3d854-bdab-451d-89a3-1d9e8fa19200 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.112703] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f08249-ad29-4d2d-967c-3f56c110f3f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.121463] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ef5812-5978-488a-b853-fb6530cc9823 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.137285] env[61947]: DEBUG nova.compute.provider_tree [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.272408] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Releasing lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.272842] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.273046] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.273391] env[61947]: DEBUG oslo_concurrency.lockutils [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] Acquired lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.273518] env[61947]: DEBUG nova.network.neutron [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Refreshing network info cache for port f599e83f-e3a3-42fc-af49-1870ab532e0e {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 623.274608] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca0656f5-305b-4581-b77d-c95ac432d762 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.286135] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312a65f7-4f1d-40c5-a8a9-baec4ebd82fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.314633] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 041b3d99-1b11-47f5-9339-2fbc25036cb7 could not be found. [ 623.314985] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 623.315187] env[61947]: INFO nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 623.315477] env[61947]: DEBUG oslo.service.loopingcall [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.315686] env[61947]: DEBUG nova.compute.manager [-] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.315776] env[61947]: DEBUG nova.network.neutron [-] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.348823] env[61947]: DEBUG nova.network.neutron [-] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.645303] env[61947]: DEBUG nova.scheduler.client.report [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.678442] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.678717] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.796153] env[61947]: DEBUG nova.network.neutron [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.851515] env[61947]: DEBUG nova.network.neutron [-] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.917557] env[61947]: DEBUG nova.network.neutron [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.152820] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.153832] env[61947]: ERROR nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Traceback (most recent call last): [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.driver.spawn(context, instance, image_meta, [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] vm_ref = self.build_virtual_machine(instance, [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.153832] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] for vif in network_info: [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self._sync_wrapper(fn, *args, **kwargs) [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.wait() [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self[:] = self._gt.wait() [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self._exit_event.wait() [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] result = hub.switch() [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.154256] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return self.greenlet.switch() [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] result = function(*args, **kwargs) [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] return func(*args, **kwargs) [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise e [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] nwinfo = self.network_api.allocate_for_instance( [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] created_port_ids = self._update_ports_for_instance( [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] with excutils.save_and_reraise_exception(): [ 624.154936] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] self.force_reraise() [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise self.value [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] updated_port = self._update_port( [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] _ensure_no_port_binding_failure(port) [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] raise exception.PortBindingFailed(port_id=port['id']) [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] nova.exception.PortBindingFailed: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. [ 624.156199] env[61947]: ERROR nova.compute.manager [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] [ 624.158149] env[61947]: DEBUG nova.compute.utils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.158149] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.022s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.158149] env[61947]: INFO nova.compute.claims [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.161017] env[61947]: DEBUG nova.compute.manager [req-2687cd9f-90b9-4a83-816e-64975c8f6024 req-cb317e54-2c96-46bf-9f3f-3c5917ff8ebd service nova] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Received event network-vif-deleted-f599e83f-e3a3-42fc-af49-1870ab532e0e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.161790] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Build of instance 5060a8b3-2492-4da9-a90d-9a2e5d502092 was re-scheduled: Binding failed for port 7f105d0f-7633-43f6-a03d-4033dbb579bf, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.162900] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.162900] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquiring lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.162900] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Acquired lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.162900] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.357097] env[61947]: INFO nova.compute.manager [-] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Took 1.04 seconds to deallocate network for instance. [ 624.358138] env[61947]: DEBUG nova.compute.claims [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 624.358538] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.422153] env[61947]: DEBUG oslo_concurrency.lockutils [req-13a885aa-f191-456e-ab82-243b289e3410 req-4f3a01d1-3924-4e88-850e-467e90853dda service nova] Releasing lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.688837] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.797826] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.300375] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Releasing lock "refresh_cache-5060a8b3-2492-4da9-a90d-9a2e5d502092" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.300696] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.300867] env[61947]: DEBUG nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.301041] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.338064] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.631147] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cf3162-7933-4f64-b691-4f49d0fc4410 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.639759] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cb13d1-4799-4d65-9459-2c5c25d0326d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.673250] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0fc17f-b454-4555-afe5-64b45c1767cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.681923] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd0c5af-3a8d-41fb-9f1b-a0c03fc732f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.697055] env[61947]: DEBUG nova.compute.provider_tree [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.844541] env[61947]: DEBUG nova.network.neutron [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.201681] env[61947]: DEBUG nova.scheduler.client.report [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.347721] env[61947]: INFO nova.compute.manager [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] [instance: 5060a8b3-2492-4da9-a90d-9a2e5d502092] Took 1.05 seconds to deallocate network for instance. [ 626.707383] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.707935] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.710690] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.816s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.712253] env[61947]: INFO nova.compute.claims [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.217221] env[61947]: DEBUG nova.compute.utils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.224403] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.224403] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.287863] env[61947]: DEBUG nova.policy [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41d4a2d39ebc4ce0a0e76350f907e7df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58b83ea896bd4654aaa9e81f635a6a55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 627.384892] env[61947]: INFO nova.scheduler.client.report [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Deleted allocations for instance 5060a8b3-2492-4da9-a90d-9a2e5d502092 [ 627.726326] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.893562] env[61947]: DEBUG oslo_concurrency.lockutils [None req-85b894a2-0a4d-4b35-bf52-90184c536c98 tempest-ImagesOneServerNegativeTestJSON-285194761 tempest-ImagesOneServerNegativeTestJSON-285194761-project-member] Lock "5060a8b3-2492-4da9-a90d-9a2e5d502092" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.187s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.942806] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Successfully created port: 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.143957] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c484950-41b7-4843-b49c-9cdd18f35469 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.156833] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21be5fe7-40c5-4424-a010-f300ee800220 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.200074] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31317487-6733-4f00-a752-998de8bb4b38 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.207176] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a61f4a1-b284-4177-b19b-c9ffbb7e714e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.221642] env[61947]: DEBUG nova.compute.provider_tree [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.397082] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.724542] env[61947]: DEBUG nova.scheduler.client.report [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.742588] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.775455] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.775844] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.775921] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.776896] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.777606] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.777606] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.777840] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.778147] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.778689] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.778889] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.778998] env[61947]: DEBUG nova.virt.hardware [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.780602] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd2c27d-9cac-441c-bc5c-eb491f8af6d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.790344] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e82f53-3f1b-45c4-9c63-22a6834c8bbd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.932235] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.230555] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.231881] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.236196] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.990s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.236196] env[61947]: DEBUG nova.objects.instance [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lazy-loading 'resources' on Instance uuid b35cb8d5-c376-4379-8021-a1e06b09c798 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 629.660059] env[61947]: DEBUG nova.compute.manager [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Received event network-changed-0e380af5-a2ee-4fca-b172-5fe8fd69fdb6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.660324] env[61947]: DEBUG nova.compute.manager [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Refreshing instance network info cache due to event network-changed-0e380af5-a2ee-4fca-b172-5fe8fd69fdb6. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 629.660469] env[61947]: DEBUG oslo_concurrency.lockutils [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] Acquiring lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.660609] env[61947]: DEBUG oslo_concurrency.lockutils [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] Acquired lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.660774] env[61947]: DEBUG nova.network.neutron [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Refreshing network info cache for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 629.737935] env[61947]: DEBUG nova.compute.utils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.740035] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.740953] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 629.854583] env[61947]: ERROR nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 629.854583] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.854583] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.854583] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.854583] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.854583] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.854583] env[61947]: ERROR nova.compute.manager raise self.value [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.854583] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.854583] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.854583] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.855101] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.855101] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.855101] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 629.855101] env[61947]: ERROR nova.compute.manager [ 629.855101] env[61947]: Traceback (most recent call last): [ 629.855101] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.855101] env[61947]: listener.cb(fileno) [ 629.855101] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.855101] env[61947]: result = function(*args, **kwargs) [ 629.855101] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.855101] env[61947]: return func(*args, **kwargs) [ 629.855101] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.855101] env[61947]: raise e [ 629.855101] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.855101] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 629.855101] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.855101] env[61947]: created_port_ids = self._update_ports_for_instance( [ 629.855101] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.855101] env[61947]: with excutils.save_and_reraise_exception(): [ 629.855101] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.855101] env[61947]: self.force_reraise() [ 629.855101] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.855101] env[61947]: raise self.value [ 629.855101] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.855101] env[61947]: updated_port = self._update_port( [ 629.855101] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.855101] env[61947]: _ensure_no_port_binding_failure(port) [ 629.855101] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.855101] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.856122] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 629.856122] env[61947]: Removing descriptor: 17 [ 629.856122] env[61947]: ERROR nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Traceback (most recent call last): [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] yield resources [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.driver.spawn(context, instance, image_meta, [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.856122] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] vm_ref = self.build_virtual_machine(instance, [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] for vif in network_info: [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self._sync_wrapper(fn, *args, **kwargs) [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.wait() [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self[:] = self._gt.wait() [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self._exit_event.wait() [ 629.856679] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] result = hub.switch() [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self.greenlet.switch() [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] result = function(*args, **kwargs) [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return func(*args, **kwargs) [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise e [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] nwinfo = self.network_api.allocate_for_instance( [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.859048] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] created_port_ids = self._update_ports_for_instance( [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] with excutils.save_and_reraise_exception(): [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.force_reraise() [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise self.value [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] updated_port = self._update_port( [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] _ensure_no_port_binding_failure(port) [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.859460] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise exception.PortBindingFailed(port_id=port['id']) [ 629.859870] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 629.859870] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] [ 629.859870] env[61947]: INFO nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Terminating instance [ 629.859870] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.863767] env[61947]: DEBUG nova.policy [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e74ccc320df4f4cb9e8a29b3bbae7bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb67ddedb2c14db39d0d0883483239e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 630.152103] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "3d4b9119-f375-45a0-99c6-ceb75f43c776" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.152103] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "3d4b9119-f375-45a0-99c6-ceb75f43c776" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.252029] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.336211] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4efbb49-5663-4e8e-ad6d-5056efa69507 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.346654] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127fdc20-48ab-4e1e-b734-de1c4aff61e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.382558] env[61947]: DEBUG nova.network.neutron [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.386784] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ed6a09-24d5-4258-be4b-bbbdf35f27f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.396817] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59eaf060-73cf-4775-9723-2165f5626516 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.418025] env[61947]: DEBUG nova.compute.provider_tree [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.596564] env[61947]: DEBUG nova.network.neutron [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.688546] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Successfully created port: 9c9ebfce-7535-45b5-8732-b7706b7fdecc {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.921430] env[61947]: DEBUG nova.scheduler.client.report [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.021361] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "46dbb96b-dc8b-4fc9-bc54-181c5a308916" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.021361] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "46dbb96b-dc8b-4fc9-bc54-181c5a308916" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.102036] env[61947]: DEBUG oslo_concurrency.lockutils [req-d815354b-cc74-4698-a30d-b810430b3d5e req-ee506788-c004-4c51-ba84-22274b40a299 service nova] Releasing lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.102036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.102036] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.261389] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.285963] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.286528] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.286816] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.287120] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.287437] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.287737] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.288063] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.288354] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.288655] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.288949] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.289264] env[61947]: DEBUG nova.virt.hardware [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.290496] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1947f3ca-d698-4634-a2e4-1cd6f03843f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.300483] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d895c4-51d2-43ba-bee6-a5d43a72e62a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.428196] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.191s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.438007] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.762s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.439892] env[61947]: INFO nova.compute.claims [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.469953] env[61947]: INFO nova.scheduler.client.report [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Deleted allocations for instance b35cb8d5-c376-4379-8021-a1e06b09c798 [ 631.627733] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.778735] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.831537] env[61947]: DEBUG nova.compute.manager [req-f4aff23c-6432-4582-8d24-cfc153a8bce7 req-f4d9a21e-8020-449a-bf99-85ad77e93f1f service nova] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Received event network-vif-deleted-0e380af5-a2ee-4fca-b172-5fe8fd69fdb6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 631.986873] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9978a27-aba9-43b5-a6d6-874ee41f6c20 tempest-ServerShowV254Test-479368687 tempest-ServerShowV254Test-479368687-project-member] Lock "b35cb8d5-c376-4379-8021-a1e06b09c798" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.642s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.175410] env[61947]: ERROR nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 632.175410] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.175410] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.175410] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.175410] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.175410] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.175410] env[61947]: ERROR nova.compute.manager raise self.value [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.175410] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.175410] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.175410] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.176247] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.176247] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.176247] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 632.176247] env[61947]: ERROR nova.compute.manager [ 632.176247] env[61947]: Traceback (most recent call last): [ 632.176247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.176247] env[61947]: listener.cb(fileno) [ 632.176247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.176247] env[61947]: result = function(*args, **kwargs) [ 632.176247] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.176247] env[61947]: return func(*args, **kwargs) [ 632.176247] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.176247] env[61947]: raise e [ 632.176247] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.176247] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 632.176247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.176247] env[61947]: created_port_ids = self._update_ports_for_instance( [ 632.176247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.176247] env[61947]: with excutils.save_and_reraise_exception(): [ 632.176247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.176247] env[61947]: self.force_reraise() [ 632.176247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.176247] env[61947]: raise self.value [ 632.176247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.176247] env[61947]: updated_port = self._update_port( [ 632.176247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.176247] env[61947]: _ensure_no_port_binding_failure(port) [ 632.176247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.176247] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.179098] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 632.179098] env[61947]: Removing descriptor: 18 [ 632.179098] env[61947]: ERROR nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] Traceback (most recent call last): [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] yield resources [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.driver.spawn(context, instance, image_meta, [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.179098] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] vm_ref = self.build_virtual_machine(instance, [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] for vif in network_info: [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self._sync_wrapper(fn, *args, **kwargs) [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.wait() [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self[:] = self._gt.wait() [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self._exit_event.wait() [ 632.179631] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] result = hub.switch() [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self.greenlet.switch() [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] result = function(*args, **kwargs) [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return func(*args, **kwargs) [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise e [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] nwinfo = self.network_api.allocate_for_instance( [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.180267] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] created_port_ids = self._update_ports_for_instance( [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] with excutils.save_and_reraise_exception(): [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.force_reraise() [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise self.value [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] updated_port = self._update_port( [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] _ensure_no_port_binding_failure(port) [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.180749] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise exception.PortBindingFailed(port_id=port['id']) [ 632.181282] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 632.181282] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] [ 632.181282] env[61947]: INFO nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Terminating instance [ 632.181282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquiring lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.181282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquired lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.181282] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.282669] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.283383] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.283637] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 632.283995] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54625341-aeb5-4059-82f4-38a8aa8ac402 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.295006] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df80b77-9bc5-4bc6-ab4b-9fb77d7c6f85 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.319883] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 77f108ae-487b-4940-bb81-b1443ccb6e55 could not be found. [ 632.320133] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 632.320308] env[61947]: INFO nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.320550] env[61947]: DEBUG oslo.service.loopingcall [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.320772] env[61947]: DEBUG nova.compute.manager [-] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.320865] env[61947]: DEBUG nova.network.neutron [-] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.335622] env[61947]: DEBUG nova.network.neutron [-] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.702931] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.792376] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.838383] env[61947]: DEBUG nova.network.neutron [-] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.926500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292c5677-6844-40b4-a28b-8b6c23532adf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.936927] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080b0baf-4766-478b-bcee-3fef250a3f91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.971039] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16e59a4-ccb2-4722-97e1-056b38926163 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.979499] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50261d1-d8d5-4bfe-b940-088d6083c292 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.994836] env[61947]: DEBUG nova.compute.provider_tree [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.295655] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Releasing lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.296325] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.296870] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.298931] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1a4ec4c-b69e-4aff-b4c2-ade3fd0a6a43 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.308031] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64de40a-4044-48aa-b27e-dc41980a43a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.333967] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e384076d-7699-450c-839f-c36a8abff553 could not be found. [ 633.334224] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.334402] env[61947]: INFO nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.334646] env[61947]: DEBUG oslo.service.loopingcall [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.334868] env[61947]: DEBUG nova.compute.manager [-] [instance: e384076d-7699-450c-839f-c36a8abff553] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.334960] env[61947]: DEBUG nova.network.neutron [-] [instance: e384076d-7699-450c-839f-c36a8abff553] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.343701] env[61947]: INFO nova.compute.manager [-] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Took 1.02 seconds to deallocate network for instance. [ 633.346986] env[61947]: DEBUG nova.compute.claims [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 633.347177] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.350474] env[61947]: DEBUG nova.network.neutron [-] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.499456] env[61947]: DEBUG nova.scheduler.client.report [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.807723] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "607d29ad-2db6-4146-a0df-192f727e9d31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.808038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.853609] env[61947]: DEBUG nova.network.neutron [-] [instance: e384076d-7699-450c-839f-c36a8abff553] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.921541] env[61947]: DEBUG nova.compute.manager [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Received event network-changed-9c9ebfce-7535-45b5-8732-b7706b7fdecc {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 633.921742] env[61947]: DEBUG nova.compute.manager [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Refreshing instance network info cache due to event network-changed-9c9ebfce-7535-45b5-8732-b7706b7fdecc. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 633.921972] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] Acquiring lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.922178] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] Acquired lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.922283] env[61947]: DEBUG nova.network.neutron [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Refreshing network info cache for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.004912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.005559] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.008542] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.813s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.357559] env[61947]: INFO nova.compute.manager [-] [instance: e384076d-7699-450c-839f-c36a8abff553] Took 1.02 seconds to deallocate network for instance. [ 634.360258] env[61947]: DEBUG nova.compute.claims [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 634.360465] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.443210] env[61947]: DEBUG nova.network.neutron [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.515559] env[61947]: DEBUG nova.compute.utils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.522936] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.523238] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.574486] env[61947]: DEBUG nova.network.neutron [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.597667] env[61947]: DEBUG nova.policy [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0aee4169c65243af802c3dcc0d27acc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96a19be770ea41a6a339316ae49570ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 634.909751] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Successfully created port: b335b945-165b-4a70-8ffa-1720bfa0dd5d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.964253] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4f5eb8-8705-483f-83ee-e57b931e0dbf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.973374] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d453c59a-dba7-4c6d-8da5-c73ae589aae1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.005062] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4378c6-c545-4966-8b45-74be261374af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.013884] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5092811-6ab4-4a78-88b4-7845c3b97ee6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.028739] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.031683] env[61947]: DEBUG nova.compute.provider_tree [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.077602] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] Releasing lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.077878] env[61947]: DEBUG nova.compute.manager [req-c3a83dd5-8226-47f9-b1ca-3a4364bb40cc req-2e99148a-fb02-4a1a-9f60-a468bd22d16c service nova] [instance: e384076d-7699-450c-839f-c36a8abff553] Received event network-vif-deleted-9c9ebfce-7535-45b5-8732-b7706b7fdecc {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.538957] env[61947]: DEBUG nova.scheduler.client.report [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.943511] env[61947]: ERROR nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 635.943511] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.943511] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.943511] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.943511] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.943511] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.943511] env[61947]: ERROR nova.compute.manager raise self.value [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.943511] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.943511] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.943511] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.944427] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.944427] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.944427] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 635.944427] env[61947]: ERROR nova.compute.manager [ 635.944427] env[61947]: Traceback (most recent call last): [ 635.944427] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.944427] env[61947]: listener.cb(fileno) [ 635.944427] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.944427] env[61947]: result = function(*args, **kwargs) [ 635.944427] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.944427] env[61947]: return func(*args, **kwargs) [ 635.944427] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.944427] env[61947]: raise e [ 635.944427] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.944427] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 635.944427] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.944427] env[61947]: created_port_ids = self._update_ports_for_instance( [ 635.944427] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.944427] env[61947]: with excutils.save_and_reraise_exception(): [ 635.944427] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.944427] env[61947]: self.force_reraise() [ 635.944427] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.944427] env[61947]: raise self.value [ 635.944427] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.944427] env[61947]: updated_port = self._update_port( [ 635.944427] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.944427] env[61947]: _ensure_no_port_binding_failure(port) [ 635.944427] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.944427] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.945438] env[61947]: nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 635.945438] env[61947]: Removing descriptor: 17 [ 635.949485] env[61947]: DEBUG nova.compute.manager [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Received event network-changed-b335b945-165b-4a70-8ffa-1720bfa0dd5d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.949626] env[61947]: DEBUG nova.compute.manager [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Refreshing instance network info cache due to event network-changed-b335b945-165b-4a70-8ffa-1720bfa0dd5d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 635.949924] env[61947]: DEBUG oslo_concurrency.lockutils [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] Acquiring lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.950098] env[61947]: DEBUG oslo_concurrency.lockutils [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] Acquired lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.950263] env[61947]: DEBUG nova.network.neutron [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Refreshing network info cache for port b335b945-165b-4a70-8ffa-1720bfa0dd5d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.045105] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.047715] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.048338] env[61947]: ERROR nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Traceback (most recent call last): [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.driver.spawn(context, instance, image_meta, [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] vm_ref = self.build_virtual_machine(instance, [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.048338] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] for vif in network_info: [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self._sync_wrapper(fn, *args, **kwargs) [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.wait() [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self[:] = self._gt.wait() [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self._exit_event.wait() [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] result = hub.switch() [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.048748] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return self.greenlet.switch() [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] result = function(*args, **kwargs) [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] return func(*args, **kwargs) [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise e [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] nwinfo = self.network_api.allocate_for_instance( [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] created_port_ids = self._update_ports_for_instance( [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] with excutils.save_and_reraise_exception(): [ 636.049206] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] self.force_reraise() [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise self.value [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] updated_port = self._update_port( [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] _ensure_no_port_binding_failure(port) [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] raise exception.PortBindingFailed(port_id=port['id']) [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] nova.exception.PortBindingFailed: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. [ 636.049692] env[61947]: ERROR nova.compute.manager [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] [ 636.050078] env[61947]: DEBUG nova.compute.utils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.050163] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.074s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.052231] env[61947]: INFO nova.compute.claims [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.055258] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Build of instance 5019b061-4e47-4c02-ab3f-fcd3e0cc3007 was re-scheduled: Binding failed for port 6c4f3643-7770-411e-ae05-1fc6a3d88453, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.055727] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.055961] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquiring lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.056130] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Acquired lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.056288] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.073525] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.073781] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.073937] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.074139] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.074288] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.074433] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.074660] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.074811] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.074971] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.075425] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.075672] env[61947]: DEBUG nova.virt.hardware [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.076825] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78c7bad-c0cf-44bc-b800-fc8e94c09401 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.086188] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3afda1-6419-42ec-a788-5ff6a9995cc3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.104866] env[61947]: ERROR nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Traceback (most recent call last): [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] yield resources [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.driver.spawn(context, instance, image_meta, [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] vm_ref = self.build_virtual_machine(instance, [ 636.104866] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] for vif in network_info: [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return self._sync_wrapper(fn, *args, **kwargs) [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.wait() [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self[:] = self._gt.wait() [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return self._exit_event.wait() [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.105325] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] current.throw(*self._exc) [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] result = function(*args, **kwargs) [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return func(*args, **kwargs) [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise e [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] nwinfo = self.network_api.allocate_for_instance( [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] created_port_ids = self._update_ports_for_instance( [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] with excutils.save_and_reraise_exception(): [ 636.105710] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.force_reraise() [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise self.value [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] updated_port = self._update_port( [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] _ensure_no_port_binding_failure(port) [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise exception.PortBindingFailed(port_id=port['id']) [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 636.106206] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] [ 636.106206] env[61947]: INFO nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Terminating instance [ 636.107852] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.468119] env[61947]: DEBUG nova.network.neutron [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.565586] env[61947]: DEBUG nova.network.neutron [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.577911] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.647177] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.067896] env[61947]: DEBUG oslo_concurrency.lockutils [req-44c7c915-125b-4bfc-8691-9c8f20f809ad req-fcd22ec4-7851-4747-8dd1-85ea740dcb8b service nova] Releasing lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.069283] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquired lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.069283] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.152189] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Releasing lock "refresh_cache-5019b061-4e47-4c02-ab3f-fcd3e0cc3007" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.152419] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.152600] env[61947]: DEBUG nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.152775] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.170866] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.504870] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1229062-2524-4826-b824-f1de484a47e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.513394] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d096e3b-e1c4-4e53-a710-769b64096c16 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.545898] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73a4aca-dfa0-48aa-8eba-ae5123be5ffc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.554967] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7d2009-daf6-4bac-b061-0378121a092b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.570412] env[61947]: DEBUG nova.compute.provider_tree [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.599028] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.675213] env[61947]: DEBUG nova.network.neutron [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.763231] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.979182] env[61947]: DEBUG nova.compute.manager [req-8494955a-60bf-4139-b4a0-3bed83a23039 req-f486b06f-4ab5-490b-b891-893c10735cf2 service nova] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Received event network-vif-deleted-b335b945-165b-4a70-8ffa-1720bfa0dd5d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.074551] env[61947]: DEBUG nova.scheduler.client.report [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.179484] env[61947]: INFO nova.compute.manager [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] [instance: 5019b061-4e47-4c02-ab3f-fcd3e0cc3007] Took 1.03 seconds to deallocate network for instance. [ 638.273030] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Releasing lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.273030] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.273181] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.273492] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-918d2602-4a2d-4707-91be-a5aadd790eb8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.283587] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a61117-2b37-41b8-b79a-a25c862bea58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.310837] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c399af8-6d17-4a74-ae32-dfa54b03a077 could not be found. [ 638.311155] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 638.311270] env[61947]: INFO nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.311518] env[61947]: DEBUG oslo.service.loopingcall [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.312208] env[61947]: DEBUG nova.compute.manager [-] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.312208] env[61947]: DEBUG nova.network.neutron [-] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.325854] env[61947]: DEBUG nova.network.neutron [-] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.581528] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.582092] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.584794] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.909s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.788349] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "be1fdafa-3c4d-430c-99c5-202e6ccebf26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.788606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "be1fdafa-3c4d-430c-99c5-202e6ccebf26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.828658] env[61947]: DEBUG nova.network.neutron [-] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.963581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "683d40c9-b738-486a-806a-7e895637ee85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.963805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.088843] env[61947]: DEBUG nova.compute.utils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.093897] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.094082] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 639.148694] env[61947]: DEBUG nova.policy [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e72a3a5ba35943b6a374abaa3e774115', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '634b11b08e6d4c2eb88d7eabe6c8e729', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 639.208367] env[61947]: INFO nova.scheduler.client.report [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Deleted allocations for instance 5019b061-4e47-4c02-ab3f-fcd3e0cc3007 [ 639.332224] env[61947]: INFO nova.compute.manager [-] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Took 1.02 seconds to deallocate network for instance. [ 639.335090] env[61947]: DEBUG nova.compute.claims [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 639.335338] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.430053] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Successfully created port: dd6cc7bb-9a65-4999-8278-96355e548097 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.489549] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ace642-f9bd-4582-b7ea-e73db1270e8e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.498435] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a274198-0328-42f1-be34-33b48173601a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.530514] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e9e3ce-7261-4afe-a677-bf23ee007873 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.538694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8761310c-ce85-42d7-a718-5f984f35c84f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.553511] env[61947]: DEBUG nova.compute.provider_tree [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.594345] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.716867] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3fa45bae-07eb-4a21-927d-87fca2bee9b8 tempest-ServersWithSpecificFlavorTestJSON-2128507702 tempest-ServersWithSpecificFlavorTestJSON-2128507702-project-member] Lock "5019b061-4e47-4c02-ab3f-fcd3e0cc3007" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.211s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.058519] env[61947]: DEBUG nova.scheduler.client.report [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.222579] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.409204] env[61947]: DEBUG nova.compute.manager [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Received event network-changed-dd6cc7bb-9a65-4999-8278-96355e548097 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 640.409407] env[61947]: DEBUG nova.compute.manager [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Refreshing instance network info cache due to event network-changed-dd6cc7bb-9a65-4999-8278-96355e548097. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 640.409617] env[61947]: DEBUG oslo_concurrency.lockutils [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] Acquiring lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.409775] env[61947]: DEBUG oslo_concurrency.lockutils [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] Acquired lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.409964] env[61947]: DEBUG nova.network.neutron [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Refreshing network info cache for port dd6cc7bb-9a65-4999-8278-96355e548097 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 640.566018] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.566018] env[61947]: ERROR nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Traceback (most recent call last): [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.driver.spawn(context, instance, image_meta, [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.566018] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] vm_ref = self.build_virtual_machine(instance, [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] for vif in network_info: [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self._sync_wrapper(fn, *args, **kwargs) [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.wait() [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self[:] = self._gt.wait() [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self._exit_event.wait() [ 640.567654] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] result = hub.switch() [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return self.greenlet.switch() [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] result = function(*args, **kwargs) [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] return func(*args, **kwargs) [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise e [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] nwinfo = self.network_api.allocate_for_instance( [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.568129] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] created_port_ids = self._update_ports_for_instance( [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] with excutils.save_and_reraise_exception(): [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] self.force_reraise() [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise self.value [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] updated_port = self._update_port( [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] _ensure_no_port_binding_failure(port) [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.568695] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] raise exception.PortBindingFailed(port_id=port['id']) [ 640.569097] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] nova.exception.PortBindingFailed: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. [ 640.569097] env[61947]: ERROR nova.compute.manager [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] [ 640.569097] env[61947]: DEBUG nova.compute.utils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.572472] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.822s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.578231] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Build of instance 35edbd0c-0634-47b9-9386-66233b4dbc7d was re-scheduled: Binding failed for port d1eb481c-21d8-4f13-a981-fb6205b17e01, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.578231] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.578231] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquiring lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.578231] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Acquired lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.578504] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.608930] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 640.632148] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.632319] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.632466] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.632646] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.632797] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.632954] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.633174] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.633331] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.633493] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.633650] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.633817] env[61947]: DEBUG nova.virt.hardware [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.634970] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd2bc13-b26b-4ded-8030-6ab2b5e11542 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.643771] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999089c3-dad6-4139-87a2-929086aa8c01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.699463] env[61947]: ERROR nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 640.699463] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.699463] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.699463] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.699463] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.699463] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.699463] env[61947]: ERROR nova.compute.manager raise self.value [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.699463] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.699463] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.699463] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.704192] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.704192] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.704192] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 640.704192] env[61947]: ERROR nova.compute.manager [ 640.704192] env[61947]: Traceback (most recent call last): [ 640.704192] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.704192] env[61947]: listener.cb(fileno) [ 640.704192] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.704192] env[61947]: result = function(*args, **kwargs) [ 640.704192] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.704192] env[61947]: return func(*args, **kwargs) [ 640.704192] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.704192] env[61947]: raise e [ 640.704192] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.704192] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 640.704192] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.704192] env[61947]: created_port_ids = self._update_ports_for_instance( [ 640.704192] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.704192] env[61947]: with excutils.save_and_reraise_exception(): [ 640.704192] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.704192] env[61947]: self.force_reraise() [ 640.704192] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.704192] env[61947]: raise self.value [ 640.704192] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.704192] env[61947]: updated_port = self._update_port( [ 640.704192] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.704192] env[61947]: _ensure_no_port_binding_failure(port) [ 640.704192] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.704192] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.705844] env[61947]: nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 640.705844] env[61947]: Removing descriptor: 18 [ 640.705844] env[61947]: ERROR nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Traceback (most recent call last): [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] yield resources [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.driver.spawn(context, instance, image_meta, [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.705844] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] vm_ref = self.build_virtual_machine(instance, [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] for vif in network_info: [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self._sync_wrapper(fn, *args, **kwargs) [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.wait() [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self[:] = self._gt.wait() [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self._exit_event.wait() [ 640.706539] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] result = hub.switch() [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self.greenlet.switch() [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] result = function(*args, **kwargs) [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return func(*args, **kwargs) [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise e [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] nwinfo = self.network_api.allocate_for_instance( [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.707257] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] created_port_ids = self._update_ports_for_instance( [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] with excutils.save_and_reraise_exception(): [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.force_reraise() [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise self.value [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] updated_port = self._update_port( [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] _ensure_no_port_binding_failure(port) [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.707690] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise exception.PortBindingFailed(port_id=port['id']) [ 640.708234] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 640.708234] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] [ 640.708234] env[61947]: INFO nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Terminating instance [ 640.708234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.744318] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.956659] env[61947]: DEBUG nova.network.neutron [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.109340] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.139454] env[61947]: DEBUG nova.network.neutron [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.216434] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.513156] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc3c443-a8a1-42a8-8925-1ab6436adccc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.521620] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1da02e1-f38b-49f4-90ad-d998ff571399 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.551540] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36411713-e5b9-4efb-94e3-e696eba8d000 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.560712] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749ce402-782a-452a-9ad1-f03ad56986f0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.575167] env[61947]: DEBUG nova.compute.provider_tree [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.641668] env[61947]: DEBUG oslo_concurrency.lockutils [req-5dd366b0-71b9-4443-8916-b566d1479996 req-6de63bab-5271-4d2c-95a7-08979d5d24c8 service nova] Releasing lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.642114] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.642303] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.719533] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Releasing lock "refresh_cache-35edbd0c-0634-47b9-9386-66233b4dbc7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.719782] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.719945] env[61947]: DEBUG nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.720122] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.760942] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.078277] env[61947]: DEBUG nova.scheduler.client.report [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.172897] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.264257] env[61947]: DEBUG nova.network.neutron [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.280229] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.434725] env[61947]: DEBUG nova.compute.manager [req-34a06295-0030-4392-8f1c-e8213f7a17a0 req-a8d73892-c227-42b5-82bc-a7d598e83848 service nova] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Received event network-vif-deleted-dd6cc7bb-9a65-4999-8278-96355e548097 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 642.584858] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.585546] env[61947]: ERROR nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Traceback (most recent call last): [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.driver.spawn(context, instance, image_meta, [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] vm_ref = self.build_virtual_machine(instance, [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.585546] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] for vif in network_info: [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self._sync_wrapper(fn, *args, **kwargs) [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.wait() [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self[:] = self._gt.wait() [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self._exit_event.wait() [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] result = hub.switch() [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.585856] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return self.greenlet.switch() [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] result = function(*args, **kwargs) [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] return func(*args, **kwargs) [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise e [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] nwinfo = self.network_api.allocate_for_instance( [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] created_port_ids = self._update_ports_for_instance( [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] with excutils.save_and_reraise_exception(): [ 642.586196] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] self.force_reraise() [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise self.value [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] updated_port = self._update_port( [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] _ensure_no_port_binding_failure(port) [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] raise exception.PortBindingFailed(port_id=port['id']) [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] nova.exception.PortBindingFailed: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. [ 642.586507] env[61947]: ERROR nova.compute.manager [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] [ 642.586775] env[61947]: DEBUG nova.compute.utils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.587515] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.362s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.589249] env[61947]: INFO nova.compute.claims [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.591762] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Build of instance 4734ece4-05d3-492f-8d1a-2c113bdd557f was re-scheduled: Binding failed for port 9d5cbd0c-256c-4648-84ea-00bcc757e332, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.592192] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.592411] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquiring lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.592554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Acquired lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.592708] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.767604] env[61947]: INFO nova.compute.manager [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] [instance: 35edbd0c-0634-47b9-9386-66233b4dbc7d] Took 1.05 seconds to deallocate network for instance. [ 642.781164] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.781704] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.781893] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 642.782182] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cce1d2dc-191a-4375-a4ec-78f59f72aebf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.791758] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad111f22-4169-480f-af49-f4dfc1f0c250 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.815309] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b7e1437-75a8-43e0-adba-ba470bed6581 could not be found. [ 642.815387] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 642.815533] env[61947]: INFO nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Took 0.03 seconds to destroy the instance on the hypervisor. [ 642.815819] env[61947]: DEBUG oslo.service.loopingcall [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.816057] env[61947]: DEBUG nova.compute.manager [-] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.816155] env[61947]: DEBUG nova.network.neutron [-] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.835748] env[61947]: DEBUG nova.network.neutron [-] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.112387] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.197491] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.337390] env[61947]: DEBUG nova.network.neutron [-] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.699128] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Releasing lock "refresh_cache-4734ece4-05d3-492f-8d1a-2c113bdd557f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.699388] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.699573] env[61947]: DEBUG nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.699718] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.725346] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.800848] env[61947]: INFO nova.scheduler.client.report [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Deleted allocations for instance 35edbd0c-0634-47b9-9386-66233b4dbc7d [ 643.842349] env[61947]: INFO nova.compute.manager [-] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Took 1.02 seconds to deallocate network for instance. [ 643.845852] env[61947]: DEBUG nova.compute.claims [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 643.845852] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.064419] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5600292-4012-49c7-a285-e55fdd18d117 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.072852] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d79fde-3b97-4390-9270-1591a95e8a7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.106823] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ed5816-d52b-4b94-8e5b-81bb57a215b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.116198] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ac2e79-581d-43fb-8520-27b874382094 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.129966] env[61947]: DEBUG nova.compute.provider_tree [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.232985] env[61947]: DEBUG nova.network.neutron [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.308186] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1cc7c262-e5b1-4f86-9a07-8ba7cdd25922 tempest-ServerDiagnosticsNegativeTest-1255789691 tempest-ServerDiagnosticsNegativeTest-1255789691-project-member] Lock "35edbd0c-0634-47b9-9386-66233b4dbc7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.979s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.633823] env[61947]: DEBUG nova.scheduler.client.report [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.735713] env[61947]: INFO nova.compute.manager [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] [instance: 4734ece4-05d3-492f-8d1a-2c113bdd557f] Took 1.04 seconds to deallocate network for instance. [ 644.812202] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.142028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.142028] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.142028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.784s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.336569] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.649455] env[61947]: DEBUG nova.compute.utils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.657767] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.657908] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.732324] env[61947]: DEBUG nova.policy [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb1ec1cea23942eeaa1566c36db27a1e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24009a57a33f44f0aa4b6f9ef4180072', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 645.765660] env[61947]: INFO nova.scheduler.client.report [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Deleted allocations for instance 4734ece4-05d3-492f-8d1a-2c113bdd557f [ 646.064147] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Successfully created port: 76bac8fc-9d59-4bde-b193-2c551ea600c3 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.094215] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86af0ece-8373-4372-9b9f-044a42d8f9a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.102405] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7db9df-98e9-4fcc-b377-2225ef263df7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.133366] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f261de-a71b-4a07-8384-faad850c7c7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.142148] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9678586-8ff5-4f52-9c5d-790bb8f30c5e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.157205] env[61947]: DEBUG nova.compute.provider_tree [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.158657] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.276652] env[61947]: DEBUG oslo_concurrency.lockutils [None req-12a72251-bdec-4f79-91cc-3b756f76f995 tempest-ServerAddressesTestJSON-194092327 tempest-ServerAddressesTestJSON-194092327-project-member] Lock "4734ece4-05d3-492f-8d1a-2c113bdd557f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.190s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.662855] env[61947]: DEBUG nova.scheduler.client.report [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.782579] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.111589] env[61947]: DEBUG nova.compute.manager [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Received event network-changed-76bac8fc-9d59-4bde-b193-2c551ea600c3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.111589] env[61947]: DEBUG nova.compute.manager [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Refreshing instance network info cache due to event network-changed-76bac8fc-9d59-4bde-b193-2c551ea600c3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 647.111589] env[61947]: DEBUG oslo_concurrency.lockutils [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] Acquiring lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.111589] env[61947]: DEBUG oslo_concurrency.lockutils [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] Acquired lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.111589] env[61947]: DEBUG nova.network.neutron [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Refreshing network info cache for port 76bac8fc-9d59-4bde-b193-2c551ea600c3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 647.170773] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.171434] env[61947]: ERROR nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Traceback (most recent call last): [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.driver.spawn(context, instance, image_meta, [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] vm_ref = self.build_virtual_machine(instance, [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.171434] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] for vif in network_info: [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self._sync_wrapper(fn, *args, **kwargs) [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.wait() [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self[:] = self._gt.wait() [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self._exit_event.wait() [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] result = hub.switch() [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.171749] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return self.greenlet.switch() [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] result = function(*args, **kwargs) [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] return func(*args, **kwargs) [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise e [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] nwinfo = self.network_api.allocate_for_instance( [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] created_port_ids = self._update_ports_for_instance( [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] with excutils.save_and_reraise_exception(): [ 647.172077] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] self.force_reraise() [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise self.value [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] updated_port = self._update_port( [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] _ensure_no_port_binding_failure(port) [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] raise exception.PortBindingFailed(port_id=port['id']) [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] nova.exception.PortBindingFailed: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. [ 647.172471] env[61947]: ERROR nova.compute.manager [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] [ 647.172748] env[61947]: DEBUG nova.compute.utils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 647.177046] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Build of instance 041b3d99-1b11-47f5-9339-2fbc25036cb7 was re-scheduled: Binding failed for port f599e83f-e3a3-42fc-af49-1870ab532e0e, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 647.177046] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 647.177046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.177046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquired lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.177299] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.177594] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.180456] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.248s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.181930] env[61947]: INFO nova.compute.claims [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.210277] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.210547] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.210734] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.210977] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.214065] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.214294] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.214588] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.214832] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.215098] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.215336] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.215583] env[61947]: DEBUG nova.virt.hardware [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.216937] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268af44d-2663-4ac9-9f69-5d9346ddcabf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.226403] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd69eca-0d0e-48dd-bc44-3dfc79bfbd2a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.317087] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.323946] env[61947]: ERROR nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 647.323946] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.323946] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.323946] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.323946] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.323946] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.323946] env[61947]: ERROR nova.compute.manager raise self.value [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.323946] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.323946] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.323946] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.324670] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.324670] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.324670] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 647.324670] env[61947]: ERROR nova.compute.manager [ 647.324670] env[61947]: Traceback (most recent call last): [ 647.324670] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.324670] env[61947]: listener.cb(fileno) [ 647.324670] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.324670] env[61947]: result = function(*args, **kwargs) [ 647.324670] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.324670] env[61947]: return func(*args, **kwargs) [ 647.324670] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.324670] env[61947]: raise e [ 647.324670] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.324670] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 647.324670] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.324670] env[61947]: created_port_ids = self._update_ports_for_instance( [ 647.324670] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.324670] env[61947]: with excutils.save_and_reraise_exception(): [ 647.324670] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.324670] env[61947]: self.force_reraise() [ 647.324670] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.324670] env[61947]: raise self.value [ 647.324670] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.324670] env[61947]: updated_port = self._update_port( [ 647.324670] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.324670] env[61947]: _ensure_no_port_binding_failure(port) [ 647.324670] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.324670] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.325386] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 647.325386] env[61947]: Removing descriptor: 17 [ 647.325386] env[61947]: ERROR nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] Traceback (most recent call last): [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] yield resources [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.driver.spawn(context, instance, image_meta, [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.325386] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] vm_ref = self.build_virtual_machine(instance, [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] for vif in network_info: [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self._sync_wrapper(fn, *args, **kwargs) [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.wait() [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self[:] = self._gt.wait() [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self._exit_event.wait() [ 647.325698] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] result = hub.switch() [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self.greenlet.switch() [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] result = function(*args, **kwargs) [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return func(*args, **kwargs) [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise e [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] nwinfo = self.network_api.allocate_for_instance( [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.326035] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] created_port_ids = self._update_ports_for_instance( [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] with excutils.save_and_reraise_exception(): [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.force_reraise() [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise self.value [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] updated_port = self._update_port( [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] _ensure_no_port_binding_failure(port) [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.326356] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise exception.PortBindingFailed(port_id=port['id']) [ 647.326662] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 647.326662] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] [ 647.326662] env[61947]: INFO nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Terminating instance [ 647.327596] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.630406] env[61947]: DEBUG nova.network.neutron [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.683576] env[61947]: DEBUG nova.network.neutron [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.699498] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.792082] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.188745] env[61947]: DEBUG oslo_concurrency.lockutils [req-2685be9b-40b1-4385-a594-9c97a30fc0fa req-71d28dc8-18f8-4c0c-91cd-477a0d89f542 service nova] Releasing lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.189142] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquired lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.189333] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.300391] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Releasing lock "refresh_cache-041b3d99-1b11-47f5-9339-2fbc25036cb7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.300639] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 648.300812] env[61947]: DEBUG nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.302035] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.326572] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.611345] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6196149c-d3bc-4e18-98ba-275153669dd6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.620076] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82e08fe-348c-488c-904c-c6e18837d8a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.653920] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c2ddf3-ca50-4fc4-b743-e6985be236b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.664128] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9269753c-5d77-498d-b015-5df4cb22c69b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.678527] env[61947]: DEBUG nova.compute.provider_tree [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.708853] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.709108] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.711285] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.767944] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.832167] env[61947]: DEBUG nova.network.neutron [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.131677] env[61947]: DEBUG nova.compute.manager [req-8b1a1d95-169d-4914-82f1-dc032115f4de req-129e3958-e668-4d41-9e0e-89441fc2eb48 service nova] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Received event network-vif-deleted-76bac8fc-9d59-4bde-b193-2c551ea600c3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.182041] env[61947]: DEBUG nova.scheduler.client.report [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.270824] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Releasing lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.271282] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.271477] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 649.271768] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d81c78b2-24a1-48a8-b5e0-8a58585dc33c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.281834] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d6dd18-469f-4bbd-8210-7e6ae843fe14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.307886] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 24c4904f-aa87-4995-813a-8a591b773532 could not be found. [ 649.308137] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 649.308322] env[61947]: INFO nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Took 0.04 seconds to destroy the instance on the hypervisor. [ 649.308670] env[61947]: DEBUG oslo.service.loopingcall [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.308888] env[61947]: DEBUG nova.compute.manager [-] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.308981] env[61947]: DEBUG nova.network.neutron [-] [instance: 24c4904f-aa87-4995-813a-8a591b773532] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.324891] env[61947]: DEBUG nova.network.neutron [-] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.334354] env[61947]: INFO nova.compute.manager [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 041b3d99-1b11-47f5-9339-2fbc25036cb7] Took 1.03 seconds to deallocate network for instance. [ 649.690032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.690032] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.691044] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.344s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.827711] env[61947]: DEBUG nova.network.neutron [-] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.197025] env[61947]: DEBUG nova.compute.utils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.200027] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.200197] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.279817] env[61947]: DEBUG nova.policy [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1a264a28e194e07a44bb478e52603fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1b46ccf47084b70965ef31d214d5783', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 650.334394] env[61947]: INFO nova.compute.manager [-] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Took 1.03 seconds to deallocate network for instance. [ 650.337365] env[61947]: DEBUG nova.compute.claims [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 650.338760] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.368159] env[61947]: INFO nova.scheduler.client.report [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Deleted allocations for instance 041b3d99-1b11-47f5-9339-2fbc25036cb7 [ 650.639915] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Successfully created port: b918f9cc-baf1-4c79-ab63-a7e3314b6e2d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.704340] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.725121] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cc8d74-f88e-49b4-9064-49e943eff4b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.733702] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf5479a-948b-48b4-b8ef-143c89dfe621 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.771557] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6b5283-8572-4eab-bf3b-aa747db8a4b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.781216] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc128e9-172d-47ab-9ba2-4230e58b088c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.797912] env[61947]: DEBUG nova.compute.provider_tree [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.878673] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32a5fbe-b48a-48f0-8c25-3ff147f1c54b tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "041b3d99-1b11-47f5-9339-2fbc25036cb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.112s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.301786] env[61947]: DEBUG nova.scheduler.client.report [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.381861] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.491493] env[61947]: DEBUG nova.compute.manager [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Received event network-changed-b918f9cc-baf1-4c79-ab63-a7e3314b6e2d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 651.492679] env[61947]: DEBUG nova.compute.manager [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Refreshing instance network info cache due to event network-changed-b918f9cc-baf1-4c79-ab63-a7e3314b6e2d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 651.494101] env[61947]: DEBUG oslo_concurrency.lockutils [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] Acquiring lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.494101] env[61947]: DEBUG oslo_concurrency.lockutils [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] Acquired lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.494310] env[61947]: DEBUG nova.network.neutron [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Refreshing network info cache for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 651.709534] env[61947]: ERROR nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 651.709534] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.709534] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.709534] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.709534] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.709534] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.709534] env[61947]: ERROR nova.compute.manager raise self.value [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.709534] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.709534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.709534] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.709991] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.709991] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.709991] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 651.709991] env[61947]: ERROR nova.compute.manager [ 651.709991] env[61947]: Traceback (most recent call last): [ 651.709991] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.709991] env[61947]: listener.cb(fileno) [ 651.709991] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.709991] env[61947]: result = function(*args, **kwargs) [ 651.709991] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.709991] env[61947]: return func(*args, **kwargs) [ 651.709991] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.709991] env[61947]: raise e [ 651.709991] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.709991] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 651.709991] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.709991] env[61947]: created_port_ids = self._update_ports_for_instance( [ 651.709991] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.709991] env[61947]: with excutils.save_and_reraise_exception(): [ 651.709991] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.709991] env[61947]: self.force_reraise() [ 651.709991] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.709991] env[61947]: raise self.value [ 651.709991] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.709991] env[61947]: updated_port = self._update_port( [ 651.709991] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.709991] env[61947]: _ensure_no_port_binding_failure(port) [ 651.709991] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.709991] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.710593] env[61947]: nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 651.710593] env[61947]: Removing descriptor: 18 [ 651.721482] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.750435] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.750671] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.750817] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.751726] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.751726] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.751726] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.751726] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.751726] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.751893] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.751986] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.752110] env[61947]: DEBUG nova.virt.hardware [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.752996] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32843588-396f-485b-b50a-2fc8419a9e4a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.762132] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73af8018-3ba1-437e-8e16-9da87e122e53 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.777142] env[61947]: ERROR nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Traceback (most recent call last): [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] yield resources [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.driver.spawn(context, instance, image_meta, [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] vm_ref = self.build_virtual_machine(instance, [ 651.777142] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] for vif in network_info: [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return self._sync_wrapper(fn, *args, **kwargs) [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.wait() [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self[:] = self._gt.wait() [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return self._exit_event.wait() [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.777423] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] current.throw(*self._exc) [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] result = function(*args, **kwargs) [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return func(*args, **kwargs) [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise e [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] nwinfo = self.network_api.allocate_for_instance( [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] created_port_ids = self._update_ports_for_instance( [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] with excutils.save_and_reraise_exception(): [ 651.777714] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.force_reraise() [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise self.value [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] updated_port = self._update_port( [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] _ensure_no_port_binding_failure(port) [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise exception.PortBindingFailed(port_id=port['id']) [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 651.777983] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] [ 651.777983] env[61947]: INFO nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Terminating instance [ 651.779554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquiring lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.806715] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.116s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.807421] env[61947]: ERROR nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Traceback (most recent call last): [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.driver.spawn(context, instance, image_meta, [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] vm_ref = self.build_virtual_machine(instance, [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.807421] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] for vif in network_info: [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self._sync_wrapper(fn, *args, **kwargs) [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.wait() [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self[:] = self._gt.wait() [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self._exit_event.wait() [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] result = hub.switch() [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.807897] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return self.greenlet.switch() [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] result = function(*args, **kwargs) [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] return func(*args, **kwargs) [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise e [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] nwinfo = self.network_api.allocate_for_instance( [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] created_port_ids = self._update_ports_for_instance( [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] with excutils.save_and_reraise_exception(): [ 651.808267] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] self.force_reraise() [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise self.value [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] updated_port = self._update_port( [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] _ensure_no_port_binding_failure(port) [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] raise exception.PortBindingFailed(port_id=port['id']) [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] nova.exception.PortBindingFailed: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. [ 651.808530] env[61947]: ERROR nova.compute.manager [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] [ 651.808768] env[61947]: DEBUG nova.compute.utils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 651.809367] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.449s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.812418] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Build of instance 77f108ae-487b-4940-bb81-b1443ccb6e55 was re-scheduled: Binding failed for port 0e380af5-a2ee-4fca-b172-5fe8fd69fdb6, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 651.812928] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 651.813180] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.813326] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.813487] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.910389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.021526] env[61947]: DEBUG nova.network.neutron [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.134122] env[61947]: DEBUG nova.network.neutron [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.336787] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.434543] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.514597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.514973] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.638420] env[61947]: DEBUG oslo_concurrency.lockutils [req-77ed1e07-d6cf-458e-96fe-0ff6c5a45386 req-8f5ea5b6-1cc4-4f50-9d15-c2e70f02efc0 service nova] Releasing lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.638844] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquired lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.639044] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.718087] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014132bc-3be4-4892-9580-c26fdcaf99da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.726583] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9155f625-648a-4ec4-a9ea-5d02de100060 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.757450] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4b68cd-efeb-4f6b-bced-75a7b97b0769 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.765416] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f917720a-41e8-4b58-ab32-fad1c9907cb7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.780730] env[61947]: DEBUG nova.compute.provider_tree [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.937716] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-77f108ae-487b-4940-bb81-b1443ccb6e55" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.937987] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 652.938195] env[61947]: DEBUG nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.938364] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.953221] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.162894] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.256944] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.285621] env[61947]: DEBUG nova.scheduler.client.report [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.456828] env[61947]: DEBUG nova.network.neutron [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.532457] env[61947]: DEBUG nova.compute.manager [req-bbec4fc4-34e3-45da-842b-2a6d16042d3b req-12fb23e9-83a0-4078-8d0c-6056a8e1e9e3 service nova] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Received event network-vif-deleted-b918f9cc-baf1-4c79-ab63-a7e3314b6e2d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 653.759973] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Releasing lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.760447] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.760643] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 653.760948] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8dfbe8ce-90ec-420a-88c2-77df796da7bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.770643] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f818af-13c5-480f-9be9-32d2a570754f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.793174] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.793739] env[61947]: ERROR nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] Traceback (most recent call last): [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.driver.spawn(context, instance, image_meta, [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] vm_ref = self.build_virtual_machine(instance, [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.793739] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] for vif in network_info: [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self._sync_wrapper(fn, *args, **kwargs) [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.wait() [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self[:] = self._gt.wait() [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self._exit_event.wait() [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] result = hub.switch() [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.794055] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return self.greenlet.switch() [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] result = function(*args, **kwargs) [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] return func(*args, **kwargs) [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise e [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] nwinfo = self.network_api.allocate_for_instance( [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] created_port_ids = self._update_ports_for_instance( [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] with excutils.save_and_reraise_exception(): [ 653.794392] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] self.force_reraise() [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise self.value [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] updated_port = self._update_port( [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] _ensure_no_port_binding_failure(port) [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] raise exception.PortBindingFailed(port_id=port['id']) [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] nova.exception.PortBindingFailed: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. [ 653.794736] env[61947]: ERROR nova.compute.manager [instance: e384076d-7699-450c-839f-c36a8abff553] [ 653.795041] env[61947]: DEBUG nova.compute.utils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.795895] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance efe14457-b8d8-454e-b0a2-fd087a92b02c could not be found. [ 653.796090] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 653.796264] env[61947]: INFO nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.796489] env[61947]: DEBUG oslo.service.loopingcall [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.796905] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Build of instance e384076d-7699-450c-839f-c36a8abff553 was re-scheduled: Binding failed for port 9c9ebfce-7535-45b5-8732-b7706b7fdecc, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.797317] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.797562] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquiring lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.797718] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Acquired lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.797875] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.798811] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.464s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.801219] env[61947]: DEBUG nova.compute.manager [-] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.801319] env[61947]: DEBUG nova.network.neutron [-] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.816653] env[61947]: DEBUG nova.network.neutron [-] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.959219] env[61947]: INFO nova.compute.manager [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 77f108ae-487b-4940-bb81-b1443ccb6e55] Took 1.02 seconds to deallocate network for instance. [ 654.128340] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 654.128855] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 654.319317] env[61947]: DEBUG nova.network.neutron [-] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.321496] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.400259] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.633829] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 654.634067] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 654.634136] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 654.657791] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47d0fe9-4d6e-4b9c-ba8f-384d993941d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.665587] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa461ed9-466c-48ae-992c-096c08f53b5c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.696121] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf35314d-0753-46c6-9a17-0fbaa10c5078 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.703631] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77f9d06-b52f-47ee-b444-5b8818b96bdd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.716499] env[61947]: DEBUG nova.compute.provider_tree [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.824678] env[61947]: INFO nova.compute.manager [-] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Took 1.02 seconds to deallocate network for instance. [ 654.826887] env[61947]: DEBUG nova.compute.claims [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 654.827078] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.903189] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Releasing lock "refresh_cache-e384076d-7699-450c-839f-c36a8abff553" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.903442] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.903631] env[61947]: DEBUG nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.903797] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.918295] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.987825] env[61947]: INFO nova.scheduler.client.report [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted allocations for instance 77f108ae-487b-4940-bb81-b1443ccb6e55 [ 655.140194] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 655.140194] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 655.140194] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 655.140194] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Didn't find any instances for network info cache update. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 655.140756] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.140931] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.141093] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.141246] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.141386] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.141528] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.141658] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 655.141797] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 655.219072] env[61947]: DEBUG nova.scheduler.client.report [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.421925] env[61947]: DEBUG nova.network.neutron [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.496195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-42a26593-49a4-4ce4-b8da-7bd81639a9d5 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "77f108ae-487b-4940-bb81-b1443ccb6e55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.086s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.647617] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.723717] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.724270] env[61947]: ERROR nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Traceback (most recent call last): [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.driver.spawn(context, instance, image_meta, [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] vm_ref = self.build_virtual_machine(instance, [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.724270] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] for vif in network_info: [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return self._sync_wrapper(fn, *args, **kwargs) [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.wait() [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self[:] = self._gt.wait() [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return self._exit_event.wait() [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] current.throw(*self._exc) [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.724544] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] result = function(*args, **kwargs) [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] return func(*args, **kwargs) [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise e [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] nwinfo = self.network_api.allocate_for_instance( [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] created_port_ids = self._update_ports_for_instance( [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] with excutils.save_and_reraise_exception(): [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] self.force_reraise() [ 655.725063] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise self.value [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] updated_port = self._update_port( [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] _ensure_no_port_binding_failure(port) [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] raise exception.PortBindingFailed(port_id=port['id']) [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] nova.exception.PortBindingFailed: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. [ 655.725567] env[61947]: ERROR nova.compute.manager [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] [ 655.725567] env[61947]: DEBUG nova.compute.utils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.726244] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.983s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.728210] env[61947]: INFO nova.compute.claims [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.730824] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Build of instance 5c399af8-6d17-4a74-ae32-dfa54b03a077 was re-scheduled: Binding failed for port b335b945-165b-4a70-8ffa-1720bfa0dd5d, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.732229] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.732229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquiring lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.732229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Acquired lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.732229] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.924667] env[61947]: INFO nova.compute.manager [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] [instance: e384076d-7699-450c-839f-c36a8abff553] Took 1.02 seconds to deallocate network for instance. [ 655.999114] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.251190] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.351573] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.444547] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.444783] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.516975] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.853941] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Releasing lock "refresh_cache-5c399af8-6d17-4a74-ae32-dfa54b03a077" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.854206] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.854368] env[61947]: DEBUG nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.854535] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.869956] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.950893] env[61947]: INFO nova.scheduler.client.report [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Deleted allocations for instance e384076d-7699-450c-839f-c36a8abff553 [ 657.077952] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf96176-0629-477f-9ddd-1c9cf1760aae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.085688] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9413dd-b685-4c12-b9d1-cc0ad58b62be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.115955] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bcf303-eac6-44ff-97ae-e73e009d85e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.123453] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f895953b-c7fc-4437-8d5e-518376ce4d15 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.136451] env[61947]: DEBUG nova.compute.provider_tree [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.372275] env[61947]: DEBUG nova.network.neutron [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.461331] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f702ee56-d16b-49ef-841b-fcf952b19108 tempest-ImagesOneServerTestJSON-1863366740 tempest-ImagesOneServerTestJSON-1863366740-project-member] Lock "e384076d-7699-450c-839f-c36a8abff553" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.043s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.640075] env[61947]: DEBUG nova.scheduler.client.report [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.875107] env[61947]: INFO nova.compute.manager [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] [instance: 5c399af8-6d17-4a74-ae32-dfa54b03a077] Took 1.02 seconds to deallocate network for instance. [ 657.963975] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.145195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.145788] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.148536] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.303s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.486825] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.653562] env[61947]: DEBUG nova.compute.utils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.658915] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.659906] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.699203] env[61947]: DEBUG nova.policy [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d637587dc8b4960975169b408d07810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c9ce5962d2649e38ed6be736406591a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 658.922842] env[61947]: INFO nova.scheduler.client.report [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Deleted allocations for instance 5c399af8-6d17-4a74-ae32-dfa54b03a077 [ 659.117933] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Successfully created port: 390319f1-eb45-407b-9c74-1e350449a557 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.130244] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be83ea6-2df3-46ca-b74b-a1d224b35207 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.137931] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1ab161-f0d3-49c0-9cbd-b1d46e88d53f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.169230] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.172521] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093ae53e-ceda-4c0d-899a-786f18d2fcef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.180272] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b877633-b256-4924-8c6f-38961ee67d41 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.193172] env[61947]: DEBUG nova.compute.provider_tree [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.434961] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a1f5d3f6-bc9c-4aef-bdca-c7164af7098e tempest-ServersAdminTestJSON-673657032 tempest-ServersAdminTestJSON-673657032-project-member] Lock "5c399af8-6d17-4a74-ae32-dfa54b03a077" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.009s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.695988] env[61947]: DEBUG nova.scheduler.client.report [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.900740] env[61947]: DEBUG nova.compute.manager [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Received event network-changed-390319f1-eb45-407b-9c74-1e350449a557 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 659.900941] env[61947]: DEBUG nova.compute.manager [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Refreshing instance network info cache due to event network-changed-390319f1-eb45-407b-9c74-1e350449a557. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 659.901365] env[61947]: DEBUG oslo_concurrency.lockutils [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] Acquiring lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.901365] env[61947]: DEBUG oslo_concurrency.lockutils [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] Acquired lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.901675] env[61947]: DEBUG nova.network.neutron [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Refreshing network info cache for port 390319f1-eb45-407b-9c74-1e350449a557 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.938141] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.156395] env[61947]: ERROR nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 660.156395] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.156395] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.156395] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.156395] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.156395] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.156395] env[61947]: ERROR nova.compute.manager raise self.value [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.156395] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.156395] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.156395] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.156959] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.156959] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.156959] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 660.156959] env[61947]: ERROR nova.compute.manager [ 660.156959] env[61947]: Traceback (most recent call last): [ 660.156959] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.156959] env[61947]: listener.cb(fileno) [ 660.156959] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.156959] env[61947]: result = function(*args, **kwargs) [ 660.156959] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.156959] env[61947]: return func(*args, **kwargs) [ 660.156959] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.156959] env[61947]: raise e [ 660.156959] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.156959] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 660.156959] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.156959] env[61947]: created_port_ids = self._update_ports_for_instance( [ 660.156959] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.156959] env[61947]: with excutils.save_and_reraise_exception(): [ 660.156959] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.156959] env[61947]: self.force_reraise() [ 660.156959] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.156959] env[61947]: raise self.value [ 660.156959] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.156959] env[61947]: updated_port = self._update_port( [ 660.156959] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.156959] env[61947]: _ensure_no_port_binding_failure(port) [ 660.156959] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.156959] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.157707] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 660.157707] env[61947]: Removing descriptor: 18 [ 660.181484] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.202032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.202525] env[61947]: ERROR nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Traceback (most recent call last): [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.driver.spawn(context, instance, image_meta, [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] vm_ref = self.build_virtual_machine(instance, [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.202525] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] for vif in network_info: [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self._sync_wrapper(fn, *args, **kwargs) [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.wait() [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self[:] = self._gt.wait() [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self._exit_event.wait() [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] result = hub.switch() [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.202860] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return self.greenlet.switch() [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] result = function(*args, **kwargs) [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] return func(*args, **kwargs) [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise e [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] nwinfo = self.network_api.allocate_for_instance( [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] created_port_ids = self._update_ports_for_instance( [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] with excutils.save_and_reraise_exception(): [ 660.203156] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] self.force_reraise() [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise self.value [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] updated_port = self._update_port( [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] _ensure_no_port_binding_failure(port) [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] raise exception.PortBindingFailed(port_id=port['id']) [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] nova.exception.PortBindingFailed: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. [ 660.203469] env[61947]: ERROR nova.compute.manager [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] [ 660.203750] env[61947]: DEBUG nova.compute.utils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.205899] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.869s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.207020] env[61947]: INFO nova.compute.claims [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.209763] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Build of instance 2b7e1437-75a8-43e0-adba-ba470bed6581 was re-scheduled: Binding failed for port dd6cc7bb-9a65-4999-8278-96355e548097, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.210230] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.210454] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.210597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.210755] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.216019] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.216019] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.216019] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.216217] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.216217] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.216217] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.216217] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.216217] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.216345] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.216345] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.216345] env[61947]: DEBUG nova.virt.hardware [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.216608] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2d33a7-951b-4d7f-85af-f4fd94d1b197 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.228626] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a314ac56-1427-4f91-ba15-ac5b4ed2847d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.245146] env[61947]: ERROR nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Traceback (most recent call last): [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] yield resources [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.driver.spawn(context, instance, image_meta, [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] vm_ref = self.build_virtual_machine(instance, [ 660.245146] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] for vif in network_info: [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return self._sync_wrapper(fn, *args, **kwargs) [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.wait() [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self[:] = self._gt.wait() [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return self._exit_event.wait() [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.245486] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] current.throw(*self._exc) [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] result = function(*args, **kwargs) [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return func(*args, **kwargs) [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise e [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] nwinfo = self.network_api.allocate_for_instance( [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] created_port_ids = self._update_ports_for_instance( [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] with excutils.save_and_reraise_exception(): [ 660.245915] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.force_reraise() [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise self.value [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] updated_port = self._update_port( [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] _ensure_no_port_binding_failure(port) [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise exception.PortBindingFailed(port_id=port['id']) [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 660.246261] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] [ 660.246261] env[61947]: INFO nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Terminating instance [ 660.247984] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.252911] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquiring lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.371369] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.007347] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "refresh_cache-2b7e1437-75a8-43e0-adba-ba470bed6581" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.007662] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.007698] env[61947]: DEBUG nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.007857] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.025407] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.035538] env[61947]: DEBUG nova.network.neutron [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.037647] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.169497] env[61947]: DEBUG nova.network.neutron [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.539716] env[61947]: DEBUG nova.network.neutron [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.674025] env[61947]: DEBUG oslo_concurrency.lockutils [req-9609211e-cad1-4f46-8d14-877556ec050d req-d4f3189c-3dc2-4a43-b80e-063debd4b62a service nova] Releasing lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.675437] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquired lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.675624] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.931355] env[61947]: DEBUG nova.compute.manager [req-0e473079-42bd-4e61-b464-c04a9c6d73dd req-2659882f-a65a-406d-88c2-6080eb82bd5f service nova] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Received event network-vif-deleted-390319f1-eb45-407b-9c74-1e350449a557 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 661.959465] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f357065f-0335-4a70-b3dc-f69322043cb8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.966956] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c6a744-6f89-4488-b8fb-6bc5a0ccda98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.008250] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594427bd-8e7d-4084-9355-ab1c0f222422 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.019483] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d699813-5186-46a3-9b24-7a62ccab5a87 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.043017] env[61947]: DEBUG nova.compute.provider_tree [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.044314] env[61947]: INFO nova.compute.manager [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2b7e1437-75a8-43e0-adba-ba470bed6581] Took 1.04 seconds to deallocate network for instance. [ 662.371709] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.514828] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.546150] env[61947]: DEBUG nova.scheduler.client.report [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.018191] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Releasing lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.019980] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.019980] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.019980] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a69bdbe3-dcc5-40de-9faf-12fb7387eae8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.029871] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce45e6e2-f08e-458c-b5e4-8599d61300a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.050997] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.846s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.051555] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.058922] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.744s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.060628] env[61947]: INFO nova.compute.claims [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.067023] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a759b902-a441-404b-9884-f08c5b8e2a71 could not be found. [ 663.067023] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.067248] env[61947]: INFO nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Took 0.05 seconds to destroy the instance on the hypervisor. [ 663.067499] env[61947]: DEBUG oslo.service.loopingcall [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.068753] env[61947]: DEBUG nova.compute.manager [-] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.068753] env[61947]: DEBUG nova.network.neutron [-] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.097224] env[61947]: INFO nova.scheduler.client.report [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Deleted allocations for instance 2b7e1437-75a8-43e0-adba-ba470bed6581 [ 663.102569] env[61947]: DEBUG nova.network.neutron [-] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.572134] env[61947]: DEBUG nova.compute.utils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.574954] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.575319] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.607231] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f021235b-e696-4195-ba1a-3e25b1ccf827 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2b7e1437-75a8-43e0-adba-ba470bed6581" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.858s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.607231] env[61947]: DEBUG nova.network.neutron [-] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.627023] env[61947]: DEBUG nova.policy [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '759fcf0ecba441d6bd7ff78b8bc87a14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ce73c6fedd54c599767c508442780a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 664.061240] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Successfully created port: fdebf6fe-c1db-4605-8c2e-5d6d06372c6d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.075956] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.114906] env[61947]: INFO nova.compute.manager [-] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Took 1.04 seconds to deallocate network for instance. [ 664.114906] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.119107] env[61947]: DEBUG nova.compute.claims [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 664.119385] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.517708] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ca5e41-bb7e-4939-8658-f8f4745d1b65 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.528459] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a73d52-adba-4b2b-873e-54564a28e23b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.561513] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d455e8-6557-410c-b3f9-b8b72994fc0b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.569606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8adfa3c1-c0d1-4800-b26b-10f34a7099f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.587437] env[61947]: DEBUG nova.compute.provider_tree [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.634887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.884134] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.884330] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.091013] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.094697] env[61947]: DEBUG nova.scheduler.client.report [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.103973] env[61947]: DEBUG nova.compute.manager [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Received event network-changed-fdebf6fe-c1db-4605-8c2e-5d6d06372c6d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.104198] env[61947]: DEBUG nova.compute.manager [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Refreshing instance network info cache due to event network-changed-fdebf6fe-c1db-4605-8c2e-5d6d06372c6d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 665.104405] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] Acquiring lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.104558] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] Acquired lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.104689] env[61947]: DEBUG nova.network.neutron [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Refreshing network info cache for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 665.133066] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.133328] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.133509] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.133729] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.133787] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.133917] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.134138] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.134416] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.134491] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.134925] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.134925] env[61947]: DEBUG nova.virt.hardware [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.135888] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f28481-2cc4-4b72-8842-7bc4effd13e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.143859] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecffe19-25d6-4b89-aae5-eb5ae618211d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.288027] env[61947]: ERROR nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 665.288027] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.288027] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.288027] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.288027] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.288027] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.288027] env[61947]: ERROR nova.compute.manager raise self.value [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.288027] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.288027] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.288027] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.288705] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.288705] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.288705] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 665.288705] env[61947]: ERROR nova.compute.manager [ 665.288705] env[61947]: Traceback (most recent call last): [ 665.288705] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.288705] env[61947]: listener.cb(fileno) [ 665.288705] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.288705] env[61947]: result = function(*args, **kwargs) [ 665.288705] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.288705] env[61947]: return func(*args, **kwargs) [ 665.288705] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.288705] env[61947]: raise e [ 665.288705] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.288705] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 665.288705] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.288705] env[61947]: created_port_ids = self._update_ports_for_instance( [ 665.288705] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.288705] env[61947]: with excutils.save_and_reraise_exception(): [ 665.288705] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.288705] env[61947]: self.force_reraise() [ 665.288705] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.288705] env[61947]: raise self.value [ 665.288705] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.288705] env[61947]: updated_port = self._update_port( [ 665.288705] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.288705] env[61947]: _ensure_no_port_binding_failure(port) [ 665.288705] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.288705] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.289566] env[61947]: nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 665.289566] env[61947]: Removing descriptor: 18 [ 665.290965] env[61947]: ERROR nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Traceback (most recent call last): [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] yield resources [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.driver.spawn(context, instance, image_meta, [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] vm_ref = self.build_virtual_machine(instance, [ 665.290965] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] for vif in network_info: [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self._sync_wrapper(fn, *args, **kwargs) [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.wait() [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self[:] = self._gt.wait() [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self._exit_event.wait() [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.291430] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] result = hub.switch() [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self.greenlet.switch() [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] result = function(*args, **kwargs) [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return func(*args, **kwargs) [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise e [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] nwinfo = self.network_api.allocate_for_instance( [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] created_port_ids = self._update_ports_for_instance( [ 665.291757] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] with excutils.save_and_reraise_exception(): [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.force_reraise() [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise self.value [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] updated_port = self._update_port( [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] _ensure_no_port_binding_failure(port) [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise exception.PortBindingFailed(port_id=port['id']) [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 665.292290] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] [ 665.292742] env[61947]: INFO nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Terminating instance [ 665.294243] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquiring lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.609735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.610323] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 665.620440] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.282s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.647825] env[61947]: DEBUG nova.network.neutron [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.755946] env[61947]: DEBUG nova.network.neutron [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.122738] env[61947]: DEBUG nova.compute.utils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.124540] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.124722] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.212537] env[61947]: DEBUG nova.policy [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec0d2d78af9d4acfa5d68fc79bdf706b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2aa2da9b8cf41e6baa067fbe787cd26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 666.261882] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc49a068-406e-486b-9aeb-c0887687c1c6 req-734a51c0-3399-4ced-90e0-4b04023a08b0 service nova] Releasing lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.263401] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquired lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.263593] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.528303] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc21ee04-79ea-4e2f-931d-c7179d4c64c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.536117] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fff275-2c2c-473f-8cde-dade2d57fc70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.566424] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affbffe4-db03-4c8c-bd0b-df5f10016eb1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.573779] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7864a8de-498e-4e17-92b3-e7765f296862 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.586544] env[61947]: DEBUG nova.compute.provider_tree [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.627360] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 666.725969] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Successfully created port: 690aafa1-5330-4392-819c-9b7ca3d138eb {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.794207] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.968630] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.089894] env[61947]: DEBUG nova.scheduler.client.report [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.343476] env[61947]: DEBUG nova.compute.manager [req-d2d3f647-3a88-484b-a49e-20e7b56011db req-46ad06ca-4064-4799-8ae2-0fa85237b5df service nova] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Received event network-vif-deleted-fdebf6fe-c1db-4605-8c2e-5d6d06372c6d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.471468] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Releasing lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.471890] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.472098] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.472404] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f355a114-2a8d-4437-906c-ba6da9d8bfd6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.484961] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a264cb9b-06c6-4f02-8304-bf5ad95d3e19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.508778] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a685616-f0ca-4de1-a78d-ecb1f40c0802 could not be found. [ 667.509141] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.509430] env[61947]: INFO nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.510149] env[61947]: DEBUG oslo.service.loopingcall [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.510489] env[61947]: DEBUG nova.compute.manager [-] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.510786] env[61947]: DEBUG nova.network.neutron [-] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.539221] env[61947]: DEBUG nova.network.neutron [-] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.596656] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.597336] env[61947]: ERROR nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] Traceback (most recent call last): [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.driver.spawn(context, instance, image_meta, [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] vm_ref = self.build_virtual_machine(instance, [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.597336] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] for vif in network_info: [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self._sync_wrapper(fn, *args, **kwargs) [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.wait() [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self[:] = self._gt.wait() [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self._exit_event.wait() [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] result = hub.switch() [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.597781] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return self.greenlet.switch() [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] result = function(*args, **kwargs) [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] return func(*args, **kwargs) [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise e [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] nwinfo = self.network_api.allocate_for_instance( [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] created_port_ids = self._update_ports_for_instance( [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] with excutils.save_and_reraise_exception(): [ 667.598082] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] self.force_reraise() [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise self.value [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] updated_port = self._update_port( [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] _ensure_no_port_binding_failure(port) [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] raise exception.PortBindingFailed(port_id=port['id']) [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] nova.exception.PortBindingFailed: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. [ 667.598434] env[61947]: ERROR nova.compute.manager [instance: 24c4904f-aa87-4995-813a-8a591b773532] [ 667.598683] env[61947]: DEBUG nova.compute.utils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.599604] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.689s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.601075] env[61947]: INFO nova.compute.claims [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.604393] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Build of instance 24c4904f-aa87-4995-813a-8a591b773532 was re-scheduled: Binding failed for port 76bac8fc-9d59-4bde-b193-2c551ea600c3, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 667.604865] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 667.605106] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquiring lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.605457] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Acquired lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.606059] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.636841] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 667.673132] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.673453] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.673631] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.673828] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.677751] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.678035] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.678295] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.678461] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.678707] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.678922] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.679114] env[61947]: DEBUG nova.virt.hardware [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.680277] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3259682f-455d-4efd-8cec-708c7060de31 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.689685] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01b00d3-a162-4173-88a1-8d186082b3e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.965318] env[61947]: ERROR nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 667.965318] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.965318] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.965318] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.965318] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.965318] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.965318] env[61947]: ERROR nova.compute.manager raise self.value [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.965318] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.965318] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.965318] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.965795] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.965795] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.965795] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 667.965795] env[61947]: ERROR nova.compute.manager [ 667.965795] env[61947]: Traceback (most recent call last): [ 667.965795] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.965795] env[61947]: listener.cb(fileno) [ 667.965795] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.965795] env[61947]: result = function(*args, **kwargs) [ 667.965795] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.965795] env[61947]: return func(*args, **kwargs) [ 667.965795] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.965795] env[61947]: raise e [ 667.965795] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.965795] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 667.965795] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.965795] env[61947]: created_port_ids = self._update_ports_for_instance( [ 667.965795] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.965795] env[61947]: with excutils.save_and_reraise_exception(): [ 667.965795] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.965795] env[61947]: self.force_reraise() [ 667.965795] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.965795] env[61947]: raise self.value [ 667.965795] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.965795] env[61947]: updated_port = self._update_port( [ 667.965795] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.965795] env[61947]: _ensure_no_port_binding_failure(port) [ 667.965795] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.965795] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.966416] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 667.966416] env[61947]: Removing descriptor: 18 [ 667.966416] env[61947]: ERROR nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Traceback (most recent call last): [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] yield resources [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.driver.spawn(context, instance, image_meta, [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.966416] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] vm_ref = self.build_virtual_machine(instance, [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] for vif in network_info: [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self._sync_wrapper(fn, *args, **kwargs) [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.wait() [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self[:] = self._gt.wait() [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self._exit_event.wait() [ 667.966683] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] result = hub.switch() [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self.greenlet.switch() [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] result = function(*args, **kwargs) [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return func(*args, **kwargs) [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise e [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] nwinfo = self.network_api.allocate_for_instance( [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.966976] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] created_port_ids = self._update_ports_for_instance( [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] with excutils.save_and_reraise_exception(): [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.force_reraise() [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise self.value [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] updated_port = self._update_port( [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] _ensure_no_port_binding_failure(port) [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.967265] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise exception.PortBindingFailed(port_id=port['id']) [ 667.967544] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 667.967544] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] [ 667.967544] env[61947]: INFO nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Terminating instance [ 667.970084] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquiring lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.970292] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquired lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.970500] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.044327] env[61947]: DEBUG nova.network.neutron [-] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.141721] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.207179] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.490541] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.551165] env[61947]: INFO nova.compute.manager [-] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Took 1.04 seconds to deallocate network for instance. [ 668.556088] env[61947]: DEBUG nova.compute.claims [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 668.556088] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.583038] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.711663] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Releasing lock "refresh_cache-24c4904f-aa87-4995-813a-8a591b773532" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.711976] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 668.712162] env[61947]: DEBUG nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.712333] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.744828] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.760086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.760086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.050042] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc793d2e-4a0a-4f12-ad69-38f88665173a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.061075] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec676183-6de5-46bf-89f1-28f070bdd8c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.096216] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Releasing lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.096794] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.097398] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 669.098014] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93f4368e-d2a5-4866-9a87-5f58d9d4ac74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.100991] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74485b8c-b421-4a9d-932b-fe1100168c9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.111953] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c360ee1-e987-4457-b5da-616c360da75a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.118698] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ac65f6-61f3-410e-8d94-892a7cca537c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.142118] env[61947]: DEBUG nova.compute.provider_tree [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.145561] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 280efc93-f913-4e4a-9ef0-fb4174d7d4cc could not be found. [ 669.145901] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 669.146194] env[61947]: INFO nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 669.146532] env[61947]: DEBUG oslo.service.loopingcall [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.148624] env[61947]: DEBUG nova.compute.manager [-] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.148624] env[61947]: DEBUG nova.network.neutron [-] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.164696] env[61947]: DEBUG nova.network.neutron [-] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.246349] env[61947]: DEBUG nova.network.neutron [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.492746] env[61947]: DEBUG nova.compute.manager [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Received event network-changed-690aafa1-5330-4392-819c-9b7ca3d138eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 669.493016] env[61947]: DEBUG nova.compute.manager [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Refreshing instance network info cache due to event network-changed-690aafa1-5330-4392-819c-9b7ca3d138eb. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 669.493319] env[61947]: DEBUG oslo_concurrency.lockutils [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] Acquiring lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.493319] env[61947]: DEBUG oslo_concurrency.lockutils [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] Acquired lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.493562] env[61947]: DEBUG nova.network.neutron [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Refreshing network info cache for port 690aafa1-5330-4392-819c-9b7ca3d138eb {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.649068] env[61947]: DEBUG nova.scheduler.client.report [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.667036] env[61947]: DEBUG nova.network.neutron [-] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.751806] env[61947]: INFO nova.compute.manager [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] [instance: 24c4904f-aa87-4995-813a-8a591b773532] Took 1.04 seconds to deallocate network for instance. [ 670.023611] env[61947]: DEBUG nova.network.neutron [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.145170] env[61947]: DEBUG nova.network.neutron [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.156018] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.156018] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.158263] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.331s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.170290] env[61947]: INFO nova.compute.manager [-] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Took 1.02 seconds to deallocate network for instance. [ 670.172694] env[61947]: DEBUG nova.compute.claims [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 670.173032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.650967] env[61947]: DEBUG oslo_concurrency.lockutils [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] Releasing lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.650967] env[61947]: DEBUG nova.compute.manager [req-459ea32d-ae17-4899-b9aa-a783f8ce5db3 req-fb154ce5-28b7-4209-9e77-6d468561734a service nova] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Received event network-vif-deleted-690aafa1-5330-4392-819c-9b7ca3d138eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 670.667130] env[61947]: DEBUG nova.compute.utils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.668569] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.668751] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.720577] env[61947]: DEBUG nova.policy [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15c390905f4540ebafd9ce494ab89049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35840e2d6f704b598d0eb957762e24e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 670.781729] env[61947]: INFO nova.scheduler.client.report [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Deleted allocations for instance 24c4904f-aa87-4995-813a-8a591b773532 [ 670.922835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.923041] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.041428] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Successfully created port: 0703b7f7-e1a2-441b-b243-00ec2c88f444 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.074257] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36109131-168b-4227-9e72-2ea15bee90bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.080359] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383bf536-e6cc-4a95-89fb-db07e30f06dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.111936] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dbd236-a547-4d5c-be8d-1c4246d0c087 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.119367] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d33d305-91bc-489c-970c-61aea7575043 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.134923] env[61947]: DEBUG nova.compute.provider_tree [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.171815] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.289347] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d57e7823-cd63-4cc9-bd6e-a6f4adc17259 tempest-DeleteServersAdminTestJSON-729881986 tempest-DeleteServersAdminTestJSON-729881986-project-member] Lock "24c4904f-aa87-4995-813a-8a591b773532" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.015s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.639051] env[61947]: DEBUG nova.scheduler.client.report [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.791925] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 671.840865] env[61947]: DEBUG nova.compute.manager [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Received event network-changed-0703b7f7-e1a2-441b-b243-00ec2c88f444 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.841078] env[61947]: DEBUG nova.compute.manager [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Refreshing instance network info cache due to event network-changed-0703b7f7-e1a2-441b-b243-00ec2c88f444. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 671.841304] env[61947]: DEBUG oslo_concurrency.lockutils [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] Acquiring lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.841481] env[61947]: DEBUG oslo_concurrency.lockutils [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] Acquired lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.841584] env[61947]: DEBUG nova.network.neutron [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Refreshing network info cache for port 0703b7f7-e1a2-441b-b243-00ec2c88f444 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 672.064125] env[61947]: ERROR nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 672.064125] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.064125] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.064125] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.064125] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.064125] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.064125] env[61947]: ERROR nova.compute.manager raise self.value [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.064125] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.064125] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.064125] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.064534] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.064534] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.064534] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 672.064534] env[61947]: ERROR nova.compute.manager [ 672.064534] env[61947]: Traceback (most recent call last): [ 672.064534] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.064534] env[61947]: listener.cb(fileno) [ 672.064534] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.064534] env[61947]: result = function(*args, **kwargs) [ 672.064534] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.064534] env[61947]: return func(*args, **kwargs) [ 672.064534] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.064534] env[61947]: raise e [ 672.064534] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.064534] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 672.064534] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.064534] env[61947]: created_port_ids = self._update_ports_for_instance( [ 672.064534] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.064534] env[61947]: with excutils.save_and_reraise_exception(): [ 672.064534] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.064534] env[61947]: self.force_reraise() [ 672.064534] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.064534] env[61947]: raise self.value [ 672.064534] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.064534] env[61947]: updated_port = self._update_port( [ 672.064534] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.064534] env[61947]: _ensure_no_port_binding_failure(port) [ 672.064534] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.064534] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.065184] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 672.065184] env[61947]: Removing descriptor: 18 [ 672.144702] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.145015] env[61947]: ERROR nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Traceback (most recent call last): [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.driver.spawn(context, instance, image_meta, [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] vm_ref = self.build_virtual_machine(instance, [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.145015] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] for vif in network_info: [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return self._sync_wrapper(fn, *args, **kwargs) [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.wait() [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self[:] = self._gt.wait() [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return self._exit_event.wait() [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] current.throw(*self._exc) [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.145370] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] result = function(*args, **kwargs) [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] return func(*args, **kwargs) [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise e [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] nwinfo = self.network_api.allocate_for_instance( [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] created_port_ids = self._update_ports_for_instance( [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] with excutils.save_and_reraise_exception(): [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] self.force_reraise() [ 672.145771] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise self.value [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] updated_port = self._update_port( [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] _ensure_no_port_binding_failure(port) [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] raise exception.PortBindingFailed(port_id=port['id']) [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] nova.exception.PortBindingFailed: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. [ 672.146068] env[61947]: ERROR nova.compute.manager [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] [ 672.146068] env[61947]: DEBUG nova.compute.utils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.147250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.500s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.147431] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.147719] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 672.148276] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.631s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.150832] env[61947]: INFO nova.compute.claims [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.154899] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Build of instance efe14457-b8d8-454e-b0a2-fd087a92b02c was re-scheduled: Binding failed for port b918f9cc-baf1-4c79-ab63-a7e3314b6e2d, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.155387] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.155646] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquiring lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.155797] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Acquired lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.155963] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.157673] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305d9d83-e4bc-485a-a112-92d75b4cd727 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.170140] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e02bb4-3fca-43dd-9b81-0186860e638d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.184907] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.187790] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdbd93d-91d2-4f3d-87a5-49f27b364a93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.195544] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ae13eb-0aa0-4d93-a0c7-300d3dc45c37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.226658] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181456MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 672.227139] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.235444] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.235444] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.235603] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.235789] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.235932] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.236436] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.236436] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.236436] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.236594] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.236749] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.236913] env[61947]: DEBUG nova.virt.hardware [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.238131] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8262204-dd49-491b-b707-8f80484de651 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.246049] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d701c317-e3fb-427b-8108-f215f9d3e42c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.260394] env[61947]: ERROR nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Traceback (most recent call last): [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] yield resources [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.driver.spawn(context, instance, image_meta, [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] vm_ref = self.build_virtual_machine(instance, [ 672.260394] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] for vif in network_info: [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return self._sync_wrapper(fn, *args, **kwargs) [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.wait() [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self[:] = self._gt.wait() [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return self._exit_event.wait() [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.260711] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] current.throw(*self._exc) [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] result = function(*args, **kwargs) [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return func(*args, **kwargs) [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise e [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] nwinfo = self.network_api.allocate_for_instance( [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] created_port_ids = self._update_ports_for_instance( [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] with excutils.save_and_reraise_exception(): [ 672.261022] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.force_reraise() [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise self.value [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] updated_port = self._update_port( [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] _ensure_no_port_binding_failure(port) [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise exception.PortBindingFailed(port_id=port['id']) [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 672.261320] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] [ 672.261320] env[61947]: INFO nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Terminating instance [ 672.262668] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquiring lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.317672] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.359513] env[61947]: DEBUG nova.network.neutron [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.457049] env[61947]: DEBUG nova.network.neutron [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.682072] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.785721] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.959668] env[61947]: DEBUG oslo_concurrency.lockutils [req-ee9a23cc-c139-4227-8c06-10369ec17698 req-c233f78e-500e-4dfd-8295-43d3c9cb0bdf service nova] Releasing lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.960151] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquired lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.960370] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.290274] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Releasing lock "refresh_cache-efe14457-b8d8-454e-b0a2-fd087a92b02c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.290501] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.290701] env[61947]: DEBUG nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.290830] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.310166] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.479417] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.520066] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b928c40-bd15-4b50-8fa0-5aa941de4519 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.528074] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4063c726-aafe-4f41-a792-b26fe1d7f663 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.560459] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3613d6-80c8-4a59-ae6a-783d2f3cbb79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.567606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413aa4b9-83f9-4017-9888-fac62e450cd0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.580645] env[61947]: DEBUG nova.compute.provider_tree [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.611536] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.812837] env[61947]: DEBUG nova.network.neutron [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.906779] env[61947]: DEBUG nova.compute.manager [req-dfe00a42-bc18-4e9e-9906-0dcb9ad90ad0 req-48c81f09-51b0-417d-9537-c12397816579 service nova] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Received event network-vif-deleted-0703b7f7-e1a2-441b-b243-00ec2c88f444 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 674.085573] env[61947]: DEBUG nova.scheduler.client.report [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.114837] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Releasing lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.115319] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.115529] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 674.115822] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c762bad1-689e-4d24-ab80-5bea65fad808 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.124646] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf95feae-efff-4d2f-b2a9-9df10d147bce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.146417] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8634de70-f66d-4f54-a13b-226737cb6d4a could not be found. [ 674.147023] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 674.147023] env[61947]: INFO nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 674.147023] env[61947]: DEBUG oslo.service.loopingcall [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.147255] env[61947]: DEBUG nova.compute.manager [-] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.147306] env[61947]: DEBUG nova.network.neutron [-] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.163093] env[61947]: DEBUG nova.network.neutron [-] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.318383] env[61947]: INFO nova.compute.manager [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] [instance: efe14457-b8d8-454e-b0a2-fd087a92b02c] Took 1.03 seconds to deallocate network for instance. [ 674.590340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.590880] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.593789] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.107s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.595217] env[61947]: INFO nova.compute.claims [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.665500] env[61947]: DEBUG nova.network.neutron [-] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.101808] env[61947]: DEBUG nova.compute.utils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.104847] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.105032] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 675.153098] env[61947]: DEBUG nova.policy [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50369315f4af4151957dadfc9d71caec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2b165a970e345d1a40508983cc755aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 675.169212] env[61947]: INFO nova.compute.manager [-] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Took 1.02 seconds to deallocate network for instance. [ 675.171521] env[61947]: DEBUG nova.compute.claims [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 675.171706] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.353117] env[61947]: INFO nova.scheduler.client.report [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Deleted allocations for instance efe14457-b8d8-454e-b0a2-fd087a92b02c [ 675.478205] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Successfully created port: 3e70b32f-6c0a-46f4-85af-09276b167ec3 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.608522] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.861448] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7919d36-b2ae-4e81-bc84-59caa8a39778 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213 tempest-FloatingIPsAssociationNegativeTestJSON-1723768213-project-member] Lock "efe14457-b8d8-454e-b0a2-fd087a92b02c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.591s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.067898] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb48e86-8a04-47e0-89f1-29590983d6a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.076267] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37db48d-d79a-4fee-8e84-47087eb5ad26 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.109211] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e39a849-a3b1-4098-ba0c-bad4c972d3a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.118392] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704966fd-4084-41a0-9a44-602b09a4ff23 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.136273] env[61947]: DEBUG nova.compute.provider_tree [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.367796] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.563335] env[61947]: DEBUG nova.compute.manager [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Received event network-changed-3e70b32f-6c0a-46f4-85af-09276b167ec3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 676.563418] env[61947]: DEBUG nova.compute.manager [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Refreshing instance network info cache due to event network-changed-3e70b32f-6c0a-46f4-85af-09276b167ec3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 676.564685] env[61947]: DEBUG oslo_concurrency.lockutils [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] Acquiring lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.564685] env[61947]: DEBUG oslo_concurrency.lockutils [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] Acquired lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.564685] env[61947]: DEBUG nova.network.neutron [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Refreshing network info cache for port 3e70b32f-6c0a-46f4-85af-09276b167ec3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 676.639829] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 676.648289] env[61947]: DEBUG nova.scheduler.client.report [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.687888] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.688745] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.688745] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.688745] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.688880] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.688911] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.689218] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.689474] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.689707] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.690755] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.690755] env[61947]: DEBUG nova.virt.hardware [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.692122] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c612bb-69c0-43a6-9037-81fadcc11fc9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.699477] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1adaecd-e44b-49c0-8bee-bc57acddb23d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.741401] env[61947]: ERROR nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 676.741401] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.741401] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.741401] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.741401] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.741401] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.741401] env[61947]: ERROR nova.compute.manager raise self.value [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.741401] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 676.741401] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.741401] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 676.742450] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.742450] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 676.742450] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 676.742450] env[61947]: ERROR nova.compute.manager [ 676.742450] env[61947]: Traceback (most recent call last): [ 676.742450] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 676.742450] env[61947]: listener.cb(fileno) [ 676.742450] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.742450] env[61947]: result = function(*args, **kwargs) [ 676.742450] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.742450] env[61947]: return func(*args, **kwargs) [ 676.742450] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.742450] env[61947]: raise e [ 676.742450] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.742450] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 676.742450] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.742450] env[61947]: created_port_ids = self._update_ports_for_instance( [ 676.742450] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.742450] env[61947]: with excutils.save_and_reraise_exception(): [ 676.742450] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.742450] env[61947]: self.force_reraise() [ 676.742450] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.742450] env[61947]: raise self.value [ 676.742450] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.742450] env[61947]: updated_port = self._update_port( [ 676.742450] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.742450] env[61947]: _ensure_no_port_binding_failure(port) [ 676.742450] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.742450] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.743666] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 676.743666] env[61947]: Removing descriptor: 18 [ 676.743666] env[61947]: ERROR nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Traceback (most recent call last): [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] yield resources [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.driver.spawn(context, instance, image_meta, [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.743666] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] vm_ref = self.build_virtual_machine(instance, [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] for vif in network_info: [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self._sync_wrapper(fn, *args, **kwargs) [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.wait() [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self[:] = self._gt.wait() [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self._exit_event.wait() [ 676.744566] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] result = hub.switch() [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self.greenlet.switch() [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] result = function(*args, **kwargs) [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return func(*args, **kwargs) [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise e [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] nwinfo = self.network_api.allocate_for_instance( [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.745887] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] created_port_ids = self._update_ports_for_instance( [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] with excutils.save_and_reraise_exception(): [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.force_reraise() [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise self.value [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] updated_port = self._update_port( [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] _ensure_no_port_binding_failure(port) [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.746480] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise exception.PortBindingFailed(port_id=port['id']) [ 676.747149] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 676.747149] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] [ 676.747149] env[61947]: INFO nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Terminating instance [ 676.747149] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquiring lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.900855] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.090482] env[61947]: DEBUG nova.network.neutron [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.151524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.152055] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.155261] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.130s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.157374] env[61947]: INFO nova.compute.claims [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.224889] env[61947]: DEBUG nova.network.neutron [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.662758] env[61947]: DEBUG nova.compute.utils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.667176] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 677.667519] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 677.718994] env[61947]: DEBUG nova.policy [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65b359bb26c546628af35fddbfcc2d89', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f992ac12ce5244499efac2693a8a517d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 677.726941] env[61947]: DEBUG oslo_concurrency.lockutils [req-9c6ec502-169c-48a3-a08f-a7bbd3139524 req-7977e308-1aec-44c9-b4a8-176610bf181c service nova] Releasing lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.727743] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquired lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.727985] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.167800] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.284791] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.297935] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Successfully created port: ac3204b8-5723-41cf-9118-2fc08248828c {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.446823] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.649469] env[61947]: DEBUG nova.compute.manager [req-3524df08-e652-45c4-9d11-aae888614ae0 req-7d01fa30-ac53-4f33-963e-25acc08f2430 service nova] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Received event network-vif-deleted-3e70b32f-6c0a-46f4-85af-09276b167ec3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 678.652014] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a029454-17b0-47f1-81d1-4db74457c0af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.663686] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b4f849-aa29-4c58-83e0-b99a69b085ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.702632] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95df1233-4f95-48db-8f76-f9d8bfaa5bf2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.710947] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f544b179-8208-4938-a8db-98e739db80bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.726404] env[61947]: DEBUG nova.compute.provider_tree [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.955634] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Releasing lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.955714] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.955894] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 678.956218] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02034bbf-a002-40de-8921-5aea815d1fcf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.965193] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77eac366-b1cc-4942-b605-e84bc848efcf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.990376] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b33f5bbe-fea2-40e1-bdcc-5c4788568e7d could not be found. [ 678.990750] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 678.990927] env[61947]: INFO nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 678.991196] env[61947]: DEBUG oslo.service.loopingcall [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.991466] env[61947]: DEBUG nova.compute.manager [-] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.991600] env[61947]: DEBUG nova.network.neutron [-] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.020628] env[61947]: DEBUG nova.network.neutron [-] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.209018] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.233114] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.233114] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.233114] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.233304] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.233304] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.233304] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.233482] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.233783] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.234130] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.234472] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.234755] env[61947]: DEBUG nova.virt.hardware [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.235645] env[61947]: DEBUG nova.scheduler.client.report [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.240347] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acd5178-3149-426e-9df2-da95074ac4c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.249533] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9922fa-446e-4ed5-b9e9-6e736f1fb77f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.525640] env[61947]: DEBUG nova.network.neutron [-] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.719054] env[61947]: ERROR nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 679.719054] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.719054] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.719054] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.719054] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.719054] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.719054] env[61947]: ERROR nova.compute.manager raise self.value [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.719054] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.719054] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.719054] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.719459] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.719459] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.719459] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 679.719459] env[61947]: ERROR nova.compute.manager [ 679.719459] env[61947]: Traceback (most recent call last): [ 679.719459] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.719459] env[61947]: listener.cb(fileno) [ 679.719459] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.719459] env[61947]: result = function(*args, **kwargs) [ 679.719459] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.719459] env[61947]: return func(*args, **kwargs) [ 679.719459] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.719459] env[61947]: raise e [ 679.719459] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.719459] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 679.719459] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.719459] env[61947]: created_port_ids = self._update_ports_for_instance( [ 679.719459] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.719459] env[61947]: with excutils.save_and_reraise_exception(): [ 679.719459] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.719459] env[61947]: self.force_reraise() [ 679.719459] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.719459] env[61947]: raise self.value [ 679.719459] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.719459] env[61947]: updated_port = self._update_port( [ 679.719459] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.719459] env[61947]: _ensure_no_port_binding_failure(port) [ 679.719459] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.719459] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.720171] env[61947]: nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 679.720171] env[61947]: Removing descriptor: 18 [ 679.720171] env[61947]: ERROR nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Traceback (most recent call last): [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] yield resources [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.driver.spawn(context, instance, image_meta, [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.720171] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] vm_ref = self.build_virtual_machine(instance, [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] for vif in network_info: [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self._sync_wrapper(fn, *args, **kwargs) [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.wait() [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self[:] = self._gt.wait() [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self._exit_event.wait() [ 679.720462] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] result = hub.switch() [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self.greenlet.switch() [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] result = function(*args, **kwargs) [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return func(*args, **kwargs) [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise e [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] nwinfo = self.network_api.allocate_for_instance( [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.720761] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] created_port_ids = self._update_ports_for_instance( [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] with excutils.save_and_reraise_exception(): [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.force_reraise() [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise self.value [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] updated_port = self._update_port( [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] _ensure_no_port_binding_failure(port) [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.721064] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise exception.PortBindingFailed(port_id=port['id']) [ 679.721343] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 679.721343] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] [ 679.721343] env[61947]: INFO nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Terminating instance [ 679.722249] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.722512] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.722938] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.748340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.748340] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.748340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.628s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.034160] env[61947]: INFO nova.compute.manager [-] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Took 1.04 seconds to deallocate network for instance. [ 680.036686] env[61947]: DEBUG nova.compute.claims [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 680.037071] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.242432] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.256468] env[61947]: DEBUG nova.compute.utils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.261759] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.261759] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.320219] env[61947]: DEBUG nova.policy [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 680.394566] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.675282] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2143edc1-d668-4fae-a451-7668de1bf251 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.679036] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Successfully created port: 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.686928] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68ef381-9b5a-4a76-b2e0-f51ceec17ce3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.726104] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2fb64a-baf8-4e8d-a71e-73b0bb63f069 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.730097] env[61947]: DEBUG nova.compute.manager [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Received event network-changed-ac3204b8-5723-41cf-9118-2fc08248828c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 680.730298] env[61947]: DEBUG nova.compute.manager [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Refreshing instance network info cache due to event network-changed-ac3204b8-5723-41cf-9118-2fc08248828c. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 680.730491] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] Acquiring lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.737279] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04757f06-994b-482f-92fc-5a462c74b892 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.753281] env[61947]: DEBUG nova.compute.provider_tree [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.762123] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.899303] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.899748] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 680.899942] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 680.900267] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] Acquired lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.900440] env[61947]: DEBUG nova.network.neutron [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Refreshing network info cache for port ac3204b8-5723-41cf-9118-2fc08248828c {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.901526] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-948026b2-6c44-4251-8690-fc9bfdec34a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.915551] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e09969-5d00-4662-b2d3-5c4f044915c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.940423] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ca2d77f-a62e-4eff-bf15-7855c94bb649 could not be found. [ 680.940751] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 680.940884] env[61947]: INFO nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Took 0.04 seconds to destroy the instance on the hypervisor. [ 680.941152] env[61947]: DEBUG oslo.service.loopingcall [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.941402] env[61947]: DEBUG nova.compute.manager [-] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.941464] env[61947]: DEBUG nova.network.neutron [-] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.964313] env[61947]: DEBUG nova.network.neutron [-] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.255553] env[61947]: DEBUG nova.scheduler.client.report [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.433028] env[61947]: DEBUG nova.network.neutron [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.467696] env[61947]: DEBUG nova.network.neutron [-] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.592977] env[61947]: DEBUG nova.network.neutron [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.761735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.762403] env[61947]: ERROR nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Traceback (most recent call last): [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.driver.spawn(context, instance, image_meta, [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] vm_ref = self.build_virtual_machine(instance, [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.762403] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] for vif in network_info: [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return self._sync_wrapper(fn, *args, **kwargs) [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.wait() [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self[:] = self._gt.wait() [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return self._exit_event.wait() [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] current.throw(*self._exc) [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.762689] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] result = function(*args, **kwargs) [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] return func(*args, **kwargs) [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise e [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] nwinfo = self.network_api.allocate_for_instance( [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] created_port_ids = self._update_ports_for_instance( [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] with excutils.save_and_reraise_exception(): [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] self.force_reraise() [ 681.762990] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise self.value [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] updated_port = self._update_port( [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] _ensure_no_port_binding_failure(port) [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] raise exception.PortBindingFailed(port_id=port['id']) [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] nova.exception.PortBindingFailed: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. [ 681.763313] env[61947]: ERROR nova.compute.manager [instance: a759b902-a441-404b-9884-f08c5b8e2a71] [ 681.763313] env[61947]: DEBUG nova.compute.utils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.765402] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Build of instance a759b902-a441-404b-9884-f08c5b8e2a71 was re-scheduled: Binding failed for port 390319f1-eb45-407b-9c74-1e350449a557, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.766308] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.766723] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquiring lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.766887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Acquired lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.767059] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.768423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.134s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.773018] env[61947]: INFO nova.compute.claims [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.774748] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.815657] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.815953] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.816178] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.816302] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.816447] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.816592] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.816870] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.816965] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.817831] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.817831] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.817831] env[61947]: DEBUG nova.virt.hardware [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.818564] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ecc9a0-0ef5-49d6-95b5-539bb594d888 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.827639] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7a2926-0fb5-4053-8ec7-2225888f8f1e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.907167] env[61947]: ERROR nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 681.907167] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.907167] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.907167] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.907167] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.907167] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.907167] env[61947]: ERROR nova.compute.manager raise self.value [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.907167] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.907167] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.907167] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.907703] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.907703] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.907703] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 681.907703] env[61947]: ERROR nova.compute.manager [ 681.907703] env[61947]: Traceback (most recent call last): [ 681.907703] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.907703] env[61947]: listener.cb(fileno) [ 681.907703] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.907703] env[61947]: result = function(*args, **kwargs) [ 681.907703] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.907703] env[61947]: return func(*args, **kwargs) [ 681.907703] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.907703] env[61947]: raise e [ 681.907703] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.907703] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 681.907703] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.907703] env[61947]: created_port_ids = self._update_ports_for_instance( [ 681.907703] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.907703] env[61947]: with excutils.save_and_reraise_exception(): [ 681.907703] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.907703] env[61947]: self.force_reraise() [ 681.907703] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.907703] env[61947]: raise self.value [ 681.907703] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.907703] env[61947]: updated_port = self._update_port( [ 681.907703] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.907703] env[61947]: _ensure_no_port_binding_failure(port) [ 681.907703] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.907703] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.908430] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 681.908430] env[61947]: Removing descriptor: 17 [ 681.908430] env[61947]: ERROR nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] Traceback (most recent call last): [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] yield resources [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.driver.spawn(context, instance, image_meta, [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.908430] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] vm_ref = self.build_virtual_machine(instance, [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] for vif in network_info: [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self._sync_wrapper(fn, *args, **kwargs) [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.wait() [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self[:] = self._gt.wait() [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self._exit_event.wait() [ 681.908726] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] result = hub.switch() [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self.greenlet.switch() [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] result = function(*args, **kwargs) [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return func(*args, **kwargs) [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise e [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] nwinfo = self.network_api.allocate_for_instance( [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.909042] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] created_port_ids = self._update_ports_for_instance( [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] with excutils.save_and_reraise_exception(): [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.force_reraise() [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise self.value [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] updated_port = self._update_port( [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] _ensure_no_port_binding_failure(port) [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.909376] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise exception.PortBindingFailed(port_id=port['id']) [ 681.909647] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 681.909647] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] [ 681.909647] env[61947]: INFO nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Terminating instance [ 681.914298] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.914298] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.914298] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.973435] env[61947]: INFO nova.compute.manager [-] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Took 1.03 seconds to deallocate network for instance. [ 681.975848] env[61947]: DEBUG nova.compute.claims [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 681.976035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.098368] env[61947]: DEBUG oslo_concurrency.lockutils [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] Releasing lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.098652] env[61947]: DEBUG nova.compute.manager [req-fdf60320-60c5-4fe6-a8e6-c3bd5e326d33 req-790fd4b3-db93-4304-9a8a-3800b27e4b29 service nova] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Received event network-vif-deleted-ac3204b8-5723-41cf-9118-2fc08248828c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 682.292060] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.382032] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.430130] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.520824] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.764032] env[61947]: DEBUG nova.compute.manager [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Received event network-changed-5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 682.764032] env[61947]: DEBUG nova.compute.manager [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Refreshing instance network info cache due to event network-changed-5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 682.764032] env[61947]: DEBUG oslo_concurrency.lockutils [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] Acquiring lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.884777] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Releasing lock "refresh_cache-a759b902-a441-404b-9884-f08c5b8e2a71" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.884777] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.884885] env[61947]: DEBUG nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.885019] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.901898] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.027596] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.028514] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.028514] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 683.028514] env[61947]: DEBUG oslo_concurrency.lockutils [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] Acquired lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.028664] env[61947]: DEBUG nova.network.neutron [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Refreshing network info cache for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 683.029842] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be8e8e0f-fa4d-431e-9258-74c9f5a072de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.041900] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fd8bfb-8baa-4be7-9d4e-99f3d1cb1e92 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.067075] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2eaab038-6c55-439a-8293-da80474d369a could not be found. [ 683.067167] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 683.067274] env[61947]: INFO nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 683.067522] env[61947]: DEBUG oslo.service.loopingcall [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.070230] env[61947]: DEBUG nova.compute.manager [-] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.070353] env[61947]: DEBUG nova.network.neutron [-] [instance: 2eaab038-6c55-439a-8293-da80474d369a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.086620] env[61947]: DEBUG nova.network.neutron [-] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.264033] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413bb75d-8fe8-4f57-b623-8b9d6ec8055f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.275044] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fab179f-85af-4cda-a5af-b4e59988e533 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.313516] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2658d915-b1d4-4528-8cf5-08813376e8d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.319273] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9df705-baf6-4b21-aea6-53e7662822d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.333683] env[61947]: DEBUG nova.compute.provider_tree [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.404951] env[61947]: DEBUG nova.network.neutron [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.550458] env[61947]: DEBUG nova.network.neutron [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.589615] env[61947]: DEBUG nova.network.neutron [-] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.639158] env[61947]: DEBUG nova.network.neutron [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.838329] env[61947]: DEBUG nova.scheduler.client.report [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.908190] env[61947]: INFO nova.compute.manager [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] [instance: a759b902-a441-404b-9884-f08c5b8e2a71] Took 1.02 seconds to deallocate network for instance. [ 684.093677] env[61947]: INFO nova.compute.manager [-] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Took 1.02 seconds to deallocate network for instance. [ 684.098987] env[61947]: DEBUG nova.compute.claims [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 684.099214] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.141970] env[61947]: DEBUG oslo_concurrency.lockutils [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] Releasing lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.142275] env[61947]: DEBUG nova.compute.manager [req-84f00755-4b0f-4394-a024-a0c731c5f960 req-fabc4339-7b5b-4233-b027-2c379956ad32 service nova] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Received event network-vif-deleted-5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.342640] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.343209] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.346377] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.790s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.851011] env[61947]: DEBUG nova.compute.utils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.855349] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 684.959601] env[61947]: INFO nova.scheduler.client.report [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Deleted allocations for instance a759b902-a441-404b-9884-f08c5b8e2a71 [ 685.356728] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.376186] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba66a06-f697-402e-82d6-1a532ed6a840 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.383725] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0820b470-a5ba-4b75-bb53-c2defe6a138e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.414360] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a747a8-1ab8-4936-b055-9a6ccb72485f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.422703] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e536bacd-f8da-432c-bf4c-0c6bfc3261dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.438570] env[61947]: DEBUG nova.compute.provider_tree [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.473961] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2507b7d4-7e66-4592-9759-266674be2a8d tempest-AttachInterfacesUnderV243Test-1989975659 tempest-AttachInterfacesUnderV243Test-1989975659-project-member] Lock "a759b902-a441-404b-9884-f08c5b8e2a71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.198s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.944937] env[61947]: DEBUG nova.scheduler.client.report [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.976143] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.366532] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.400098] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.401040] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.401253] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.401452] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.401662] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.401743] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.401954] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.402126] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.402292] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.402452] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.402621] env[61947]: DEBUG nova.virt.hardware [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.403516] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069379d4-ef52-4fb3-b641-668182243fa0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.417229] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa2b3e4-c1ee-4741-a6db-4e254c3303e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.429637] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.436424] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Creating folder: Project (c2a000aa53b74eada6c685c13e0440c8). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.436746] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3506d2bc-81c4-4672-abcb-0cd49ffd822a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.446819] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Created folder: Project (c2a000aa53b74eada6c685c13e0440c8) in parent group-v264556. [ 686.447014] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Creating folder: Instances. Parent ref: group-v264568. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.447276] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e43899c0-d7bf-42b1-884e-d475b1d03687 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.450064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.104s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.450628] env[61947]: ERROR nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Traceback (most recent call last): [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.driver.spawn(context, instance, image_meta, [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] vm_ref = self.build_virtual_machine(instance, [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.450628] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] for vif in network_info: [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self._sync_wrapper(fn, *args, **kwargs) [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.wait() [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self[:] = self._gt.wait() [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self._exit_event.wait() [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] result = hub.switch() [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.450971] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return self.greenlet.switch() [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] result = function(*args, **kwargs) [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] return func(*args, **kwargs) [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise e [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] nwinfo = self.network_api.allocate_for_instance( [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] created_port_ids = self._update_ports_for_instance( [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] with excutils.save_and_reraise_exception(): [ 686.451339] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] self.force_reraise() [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise self.value [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] updated_port = self._update_port( [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] _ensure_no_port_binding_failure(port) [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] raise exception.PortBindingFailed(port_id=port['id']) [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] nova.exception.PortBindingFailed: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. [ 686.451673] env[61947]: ERROR nova.compute.manager [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] [ 686.451985] env[61947]: DEBUG nova.compute.utils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.452505] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.279s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.461082] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Build of instance 0a685616-f0ca-4de1-a78d-ecb1f40c0802 was re-scheduled: Binding failed for port fdebf6fe-c1db-4605-8c2e-5d6d06372c6d, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.461082] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.461223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquiring lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.461879] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Acquired lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.461879] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.462468] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Created folder: Instances in parent group-v264568. [ 686.462681] env[61947]: DEBUG oslo.service.loopingcall [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.463180] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 686.463274] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3bf201d-955d-482b-b37d-7a398b0ca96a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.486523] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.486523] env[61947]: value = "task-1224166" [ 686.486523] env[61947]: _type = "Task" [ 686.486523] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.498157] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224166, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.499221] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.881844] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.882297] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.000371] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224166, 'name': CreateVM_Task, 'duration_secs': 0.276485} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.003205] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 687.003827] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.004083] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.004384] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 687.004641] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee1ea347-d021-494d-a6e0-487a19fc6219 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.009720] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 687.009720] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b07da-9eaa-a6f9-1e1f-2356167734d6" [ 687.009720] env[61947]: _type = "Task" [ 687.009720] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.020833] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b07da-9eaa-a6f9-1e1f-2356167734d6, 'name': SearchDatastore_Task} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.021142] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.021372] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 687.021892] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.021892] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.021892] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 687.022201] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baf52731-b96d-40a5-8faf-3844e454e2dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.030977] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 687.031180] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 687.031899] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4457f29d-6361-4284-a09d-91ec1fa6101d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.037117] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 687.037117] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523d49b6-4dbf-76aa-53eb-77ca6d0e87af" [ 687.037117] env[61947]: _type = "Task" [ 687.037117] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.045106] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523d49b6-4dbf-76aa-53eb-77ca6d0e87af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.138119] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.280122] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.368742] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a7973e-5243-47d5-a5fd-ed09043fc9f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.377152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e43617a-7b28-4449-9b8c-8c49d68c72ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.414136] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6fdafa-465a-4695-85c1-f211959cd02c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.421689] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3aadb0-fbab-4b76-99f0-2939a96991a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.436525] env[61947]: DEBUG nova.compute.provider_tree [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.547758] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523d49b6-4dbf-76aa-53eb-77ca6d0e87af, 'name': SearchDatastore_Task, 'duration_secs': 0.007716} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.548802] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-416ed444-0fec-42ca-84f7-9105ca7217fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.554372] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 687.554372] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af9e45-1b43-f7f2-bf57-ad05ac2fda1b" [ 687.554372] env[61947]: _type = "Task" [ 687.554372] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.563314] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af9e45-1b43-f7f2-bf57-ad05ac2fda1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.785778] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Releasing lock "refresh_cache-0a685616-f0ca-4de1-a78d-ecb1f40c0802" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.785778] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.785778] env[61947]: DEBUG nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.785778] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.811116] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.942021] env[61947]: DEBUG nova.scheduler.client.report [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.065118] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af9e45-1b43-f7f2-bf57-ad05ac2fda1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008162} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.068827] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.068827] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8/0c7a6d23-7196-4af2-9f52-6e5a09f26fd8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 688.068827] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-529bc77a-7628-4881-8e86-9c05664f99ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.076687] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 688.076687] env[61947]: value = "task-1224167" [ 688.076687] env[61947]: _type = "Task" [ 688.076687] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.080029] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.316517] env[61947]: DEBUG nova.network.neutron [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.446186] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.446860] env[61947]: ERROR nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Traceback (most recent call last): [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.driver.spawn(context, instance, image_meta, [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] vm_ref = self.build_virtual_machine(instance, [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.446860] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] for vif in network_info: [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self._sync_wrapper(fn, *args, **kwargs) [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.wait() [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self[:] = self._gt.wait() [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self._exit_event.wait() [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] result = hub.switch() [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.447262] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return self.greenlet.switch() [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] result = function(*args, **kwargs) [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] return func(*args, **kwargs) [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise e [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] nwinfo = self.network_api.allocate_for_instance( [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] created_port_ids = self._update_ports_for_instance( [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] with excutils.save_and_reraise_exception(): [ 688.447568] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] self.force_reraise() [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise self.value [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] updated_port = self._update_port( [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] _ensure_no_port_binding_failure(port) [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] raise exception.PortBindingFailed(port_id=port['id']) [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] nova.exception.PortBindingFailed: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. [ 688.447963] env[61947]: ERROR nova.compute.manager [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] [ 688.448298] env[61947]: DEBUG nova.compute.utils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.450924] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.224s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.452769] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Build of instance 280efc93-f913-4e4a-9ef0-fb4174d7d4cc was re-scheduled: Binding failed for port 690aafa1-5330-4392-819c-9b7ca3d138eb, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.453266] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.453535] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquiring lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.453633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Acquired lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.453802] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.581808] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224167, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.819499] env[61947]: INFO nova.compute.manager [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] [instance: 0a685616-f0ca-4de1-a78d-ecb1f40c0802] Took 1.04 seconds to deallocate network for instance. [ 688.977172] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.077557] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.084206] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565265} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.084315] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8/0c7a6d23-7196-4af2-9f52-6e5a09f26fd8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 689.084477] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 689.084717] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a443749-84d2-4368-9eab-808814e5891d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.091216] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 689.091216] env[61947]: value = "task-1224168" [ 689.091216] env[61947]: _type = "Task" [ 689.091216] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.099424] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224168, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.580302] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Releasing lock "refresh_cache-280efc93-f913-4e4a-9ef0-fb4174d7d4cc" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.580674] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.580748] env[61947]: DEBUG nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.580934] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.602244] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.604646] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224168, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064242} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.605124] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 689.605906] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91714b91-3b00-4bdb-972a-0fabe29362cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.631437] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8/0c7a6d23-7196-4af2-9f52-6e5a09f26fd8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 689.631437] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc67396f-deb9-47f0-bb5e-9f4ef16ee8c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.650547] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 689.650547] env[61947]: value = "task-1224169" [ 689.650547] env[61947]: _type = "Task" [ 689.650547] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.660178] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224169, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.850787] env[61947]: INFO nova.scheduler.client.report [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Deleted allocations for instance 0a685616-f0ca-4de1-a78d-ecb1f40c0802 [ 689.993050] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 0a685616-f0ca-4de1-a78d-ecb1f40c0802 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.106188] env[61947]: DEBUG nova.network.neutron [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.162306] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224169, 'name': ReconfigVM_Task, 'duration_secs': 0.272753} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.162604] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8/0c7a6d23-7196-4af2-9f52-6e5a09f26fd8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 690.163306] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80f03ee4-2862-4532-b4a5-37b417d2a0e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.169672] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 690.169672] env[61947]: value = "task-1224170" [ 690.169672] env[61947]: _type = "Task" [ 690.169672] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.177709] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224170, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.358877] env[61947]: DEBUG oslo_concurrency.lockutils [None req-165dc8a2-7a54-4f7c-bd43-6721515f2182 tempest-FloatingIPsAssociationTestJSON-488202116 tempest-FloatingIPsAssociationTestJSON-488202116-project-member] Lock "0a685616-f0ca-4de1-a78d-ecb1f40c0802" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.484s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.497364] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 280efc93-f913-4e4a-9ef0-fb4174d7d4cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.497551] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8634de70-f66d-4f54-a13b-226737cb6d4a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 690.497703] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b33f5bbe-fea2-40e1-bdcc-5c4788568e7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 690.497838] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 6ca2d77f-a62e-4eff-bf15-7855c94bb649 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 690.497956] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 2eaab038-6c55-439a-8293-da80474d369a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 690.498084] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 690.611810] env[61947]: INFO nova.compute.manager [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] [instance: 280efc93-f913-4e4a-9ef0-fb4174d7d4cc] Took 1.03 seconds to deallocate network for instance. [ 690.679915] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224170, 'name': Rename_Task, 'duration_secs': 0.134732} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.680269] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 690.680518] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26ecdec3-f0eb-44ec-a647-3ee3a471655b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.686858] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 690.686858] env[61947]: value = "task-1224171" [ 690.686858] env[61947]: _type = "Task" [ 690.686858] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.694920] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.861447] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.001008] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 32fe176d-ebc2-46a4-8e93-3b2555a681d4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.198516] env[61947]: DEBUG oslo_vmware.api [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224171, 'name': PowerOnVM_Task, 'duration_secs': 0.412086} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.199113] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 691.199317] env[61947]: INFO nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Took 4.83 seconds to spawn the instance on the hypervisor. [ 691.199586] env[61947]: DEBUG nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 691.200317] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ce49c8-7d0b-43b0-9f3c-d2052b0f643b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.397660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.505016] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d58f3c69-ae7a-48d5-a796-34898c6e704c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.646395] env[61947]: INFO nova.scheduler.client.report [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Deleted allocations for instance 280efc93-f913-4e4a-9ef0-fb4174d7d4cc [ 691.723563] env[61947]: INFO nova.compute.manager [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Took 27.11 seconds to build instance. [ 692.008113] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 947b697d-4120-491d-83ce-21681974c224 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.158943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-07d960de-cf58-4742-9193-5a7eabc07a69 tempest-ServerRescueTestJSON-203026027 tempest-ServerRescueTestJSON-203026027-project-member] Lock "280efc93-f913-4e4a-9ef0-fb4174d7d4cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.142s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.228471] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3f77cb78-470b-4bd0-adc9-724e71279be0 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.079s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.512053] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance c8c66631-1fae-4a03-810f-66d64a75b118 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.661144] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.731098] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 693.016621] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.192818] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.255615] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.521932] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8a4df3f1-67f3-404d-953f-caa6ba1b1641 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.902235] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.902527] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.028041] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8c1101f6-4263-47db-bbd6-2be773ea3380 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.532358] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 94d9244e-bf54-442d-8947-71d15d3da551 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.036794] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 30b75b32-6fde-4a81-af55-11b51f3e087d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.540053] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.043495] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 3d4b9119-f375-45a0-99c6-ceb75f43c776 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.547132] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 46dbb96b-dc8b-4fc9-bc54-181c5a308916 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.051883] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 607d29ad-2db6-4146-a0df-192f727e9d31 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.557261] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance be1fdafa-3c4d-430c-99c5-202e6ccebf26 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.060926] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 683d40c9-b738-486a-806a-7e895637ee85 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.564563] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d642bdb3-8e98-4667-9af7-71dae739e6b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.068926] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance c8f37b6f-4f0c-43db-a1ea-3c45d3956f66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.571696] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9a54ca4f-e6ec-4413-b162-fca0cd824e00 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.076051] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b7d84a56-edb6-4d1e-b7e9-294078be79d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.578581] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 2fc54703-5438-4c2a-b1e7-77431ce5177d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.082143] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.586082] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.586082] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 701.586082] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 701.880508] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1fdc0b-6545-42d2-9d2e-093175f59a65 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.888606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740ac15c-49de-4718-b8a3-953622312982 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.918263] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc0e36b-8bcc-4f40-b389-affa36bba0dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.925845] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41f3b3f-06b9-481c-a2fa-b45acd427e37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.938642] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.442913] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.948511] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 702.948709] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.498s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.949022] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.631s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.950532] env[61947]: INFO nova.compute.claims [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.280471] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ffae58-5cbd-4cd2-9d23-9bdcef6e270e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.287651] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d456fe5-784e-4c79-b1cf-89773ef5ee2f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.317569] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0490f147-3ccf-4bd4-bbe4-ea795b3f6daf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.325199] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5249fdaa-7560-43e3-8928-c9f0da847fba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.338248] env[61947]: DEBUG nova.compute.provider_tree [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.841227] env[61947]: DEBUG nova.scheduler.client.report [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.347051] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.347371] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.350218] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.178s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.854387] env[61947]: DEBUG nova.compute.utils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.855914] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 706.175469] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac3d819-251a-461e-82d4-6f20ac16b359 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.184017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b592f2cc-2dbb-4959-bfe5-44ed4bf520b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.214997] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823463fe-656c-4fb4-9b63-e9cb3658f9e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.223070] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35da04b3-96eb-4ecd-be02-670dba6a7b0e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.236370] env[61947]: DEBUG nova.compute.provider_tree [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.360606] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.740652] env[61947]: DEBUG nova.scheduler.client.report [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.249939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.250672] env[61947]: ERROR nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Traceback (most recent call last): [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.driver.spawn(context, instance, image_meta, [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] vm_ref = self.build_virtual_machine(instance, [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.250672] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] for vif in network_info: [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return self._sync_wrapper(fn, *args, **kwargs) [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.wait() [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self[:] = self._gt.wait() [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return self._exit_event.wait() [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] current.throw(*self._exc) [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.250984] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] result = function(*args, **kwargs) [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] return func(*args, **kwargs) [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise e [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] nwinfo = self.network_api.allocate_for_instance( [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] created_port_ids = self._update_ports_for_instance( [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] with excutils.save_and_reraise_exception(): [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] self.force_reraise() [ 707.251323] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise self.value [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] updated_port = self._update_port( [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] _ensure_no_port_binding_failure(port) [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] raise exception.PortBindingFailed(port_id=port['id']) [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] nova.exception.PortBindingFailed: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. [ 707.251660] env[61947]: ERROR nova.compute.manager [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] [ 707.251660] env[61947]: DEBUG nova.compute.utils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.252660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.352s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.254130] env[61947]: INFO nova.compute.claims [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.256768] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Build of instance 8634de70-f66d-4f54-a13b-226737cb6d4a was re-scheduled: Binding failed for port 0703b7f7-e1a2-441b-b243-00ec2c88f444, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.257189] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.257410] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquiring lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.257553] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Acquired lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.257726] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.369607] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.395280] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.396027] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.396027] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.396027] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.396175] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.396333] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.396585] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.396779] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.397468] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.397468] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.397468] env[61947]: DEBUG nova.virt.hardware [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.398293] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcf3f03-d53a-44bc-970e-52b698347b06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.406497] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894a80cb-7f66-4c62-a717-400d3faf743f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.421135] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 707.425751] env[61947]: DEBUG oslo.service.loopingcall [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.426302] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 707.426513] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfe55283-3793-491c-8931-88cb79b11ba6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.442480] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 707.442480] env[61947]: value = "task-1224172" [ 707.442480] env[61947]: _type = "Task" [ 707.442480] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.449730] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224172, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.788610] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.879749] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.954261] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224172, 'name': CreateVM_Task, 'duration_secs': 0.269053} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.954455] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 707.954871] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.955046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.955389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 707.955659] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16f083b6-da19-47ed-96c7-03daf6884312 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.959992] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 707.959992] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528c0cdc-b123-33db-46a9-1d18a39c3952" [ 707.959992] env[61947]: _type = "Task" [ 707.959992] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.967199] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528c0cdc-b123-33db-46a9-1d18a39c3952, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.383024] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Releasing lock "refresh_cache-8634de70-f66d-4f54-a13b-226737cb6d4a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.383273] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.383405] env[61947]: DEBUG nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.383568] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.401575] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.471025] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528c0cdc-b123-33db-46a9-1d18a39c3952, 'name': SearchDatastore_Task, 'duration_secs': 0.010063} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.471207] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.471439] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.471666] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.471812] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.472032] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.472295] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5cb8880-0a89-4bf6-93da-48144d91427b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.480144] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.480321] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 708.481029] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a27184a-9159-48b2-a179-78207b2f78ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.485853] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 708.485853] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d5d7e6-6686-3bb9-8f0c-10ef2621f6b7" [ 708.485853] env[61947]: _type = "Task" [ 708.485853] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.496164] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d5d7e6-6686-3bb9-8f0c-10ef2621f6b7, 'name': SearchDatastore_Task, 'duration_secs': 0.007724} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.499114] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e35ebca-f0a4-42bd-a34d-6c523aefa41f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.503704] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 708.503704] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52025415-d11c-d20a-9e90-3e44aa88d510" [ 708.503704] env[61947]: _type = "Task" [ 708.503704] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.511063] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52025415-d11c-d20a-9e90-3e44aa88d510, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.669165] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4730562e-3365-4303-a63f-85d92825cd82 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.676756] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b655cf7c-3f89-4a3c-b113-827fcb63b967 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.705782] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac36c01-9e51-4cc2-88b7-d2027bd31be4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.712930] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b72ac6-3b15-4328-82e4-913bf31bfbe1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.726709] env[61947]: DEBUG nova.compute.provider_tree [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.903833] env[61947]: DEBUG nova.network.neutron [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.013611] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52025415-d11c-d20a-9e90-3e44aa88d510, 'name': SearchDatastore_Task, 'duration_secs': 0.007475} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.013854] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.014125] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 709.014372] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5189d33c-d65e-4ce9-ad8a-dc529b244c78 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.020598] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 709.020598] env[61947]: value = "task-1224173" [ 709.020598] env[61947]: _type = "Task" [ 709.020598] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.027874] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.230387] env[61947]: DEBUG nova.scheduler.client.report [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.407217] env[61947]: INFO nova.compute.manager [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] [instance: 8634de70-f66d-4f54-a13b-226737cb6d4a] Took 1.02 seconds to deallocate network for instance. [ 709.531772] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461085} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.531924] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 709.532108] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.532356] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11fda3fd-be07-4724-84f7-982ffa391571 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.541082] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 709.541082] env[61947]: value = "task-1224174" [ 709.541082] env[61947]: _type = "Task" [ 709.541082] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.545917] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.735374] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.736034] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.738918] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.702s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.048595] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072087} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.048915] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.049712] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ef45b4-42ce-4e66-b3a4-852200b778da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.068847] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.069068] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8c3d150-2904-4d60-a9e1-396ae6cb31c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.087214] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 710.087214] env[61947]: value = "task-1224175" [ 710.087214] env[61947]: _type = "Task" [ 710.087214] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.097121] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224175, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.248351] env[61947]: DEBUG nova.compute.utils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.252770] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.253012] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.311531] env[61947]: DEBUG nova.policy [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '426deb74dc554d0091a5e490ad43ee30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a8e086117fd405683fa97b4df802680', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 710.441116] env[61947]: INFO nova.scheduler.client.report [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Deleted allocations for instance 8634de70-f66d-4f54-a13b-226737cb6d4a [ 710.598388] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224175, 'name': ReconfigVM_Task, 'duration_secs': 0.279452} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.604267] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.605356] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fab88470-4f8d-4ce1-8806-f06293c73a11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.612747] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 710.612747] env[61947]: value = "task-1224176" [ 710.612747] env[61947]: _type = "Task" [ 710.612747] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.629693] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224176, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.631062] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76373cb-607d-42f2-bc11-7439a6c27663 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.639872] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a398cd0d-8788-4e38-aced-718a62979b18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.670286] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75c0a8a-a576-4fc8-abe2-fba3d4157d39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.677525] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dfe427-7601-4807-9460-b6407376b385 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.691147] env[61947]: DEBUG nova.compute.provider_tree [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.733251] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Successfully created port: 34ca4adb-5227-4735-99d9-7d7293987aec {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.753987] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.954767] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f26888db-d465-4226-8995-bf4622f7dbb4 tempest-InstanceActionsNegativeTestJSON-730454879 tempest-InstanceActionsNegativeTestJSON-730454879-project-member] Lock "8634de70-f66d-4f54-a13b-226737cb6d4a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.892s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.126610] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224176, 'name': Rename_Task, 'duration_secs': 0.144388} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.127663] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 711.127663] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93d5a611-e5f7-445e-9594-4740fbb239b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.133569] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 711.133569] env[61947]: value = "task-1224177" [ 711.133569] env[61947]: _type = "Task" [ 711.133569] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.147603] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.193979] env[61947]: DEBUG nova.scheduler.client.report [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.489323] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.541898] env[61947]: DEBUG nova.compute.manager [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Received event network-changed-34ca4adb-5227-4735-99d9-7d7293987aec {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 711.542131] env[61947]: DEBUG nova.compute.manager [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Refreshing instance network info cache due to event network-changed-34ca4adb-5227-4735-99d9-7d7293987aec. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 711.542357] env[61947]: DEBUG oslo_concurrency.lockutils [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] Acquiring lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.542630] env[61947]: DEBUG oslo_concurrency.lockutils [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] Acquired lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.542665] env[61947]: DEBUG nova.network.neutron [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Refreshing network info cache for port 34ca4adb-5227-4735-99d9-7d7293987aec {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.647300] env[61947]: DEBUG oslo_vmware.api [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224177, 'name': PowerOnVM_Task, 'duration_secs': 0.42796} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.647617] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 711.647863] env[61947]: INFO nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Took 4.28 seconds to spawn the instance on the hypervisor. [ 711.648096] env[61947]: DEBUG nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 711.648931] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b7bea2-5acf-4c33-89e4-43fae71b0a2c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.698957] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.699569] env[61947]: ERROR nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Traceback (most recent call last): [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.driver.spawn(context, instance, image_meta, [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] vm_ref = self.build_virtual_machine(instance, [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.699569] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] for vif in network_info: [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self._sync_wrapper(fn, *args, **kwargs) [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.wait() [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self[:] = self._gt.wait() [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self._exit_event.wait() [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] result = hub.switch() [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.700158] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return self.greenlet.switch() [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] result = function(*args, **kwargs) [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] return func(*args, **kwargs) [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise e [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] nwinfo = self.network_api.allocate_for_instance( [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] created_port_ids = self._update_ports_for_instance( [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] with excutils.save_and_reraise_exception(): [ 711.700581] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] self.force_reraise() [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise self.value [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] updated_port = self._update_port( [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] _ensure_no_port_binding_failure(port) [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] raise exception.PortBindingFailed(port_id=port['id']) [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] nova.exception.PortBindingFailed: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. [ 711.701282] env[61947]: ERROR nova.compute.manager [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] [ 711.701767] env[61947]: DEBUG nova.compute.utils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.701873] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.726s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.705713] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Build of instance b33f5bbe-fea2-40e1-bdcc-5c4788568e7d was re-scheduled: Binding failed for port 3e70b32f-6c0a-46f4-85af-09276b167ec3, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.706269] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.706623] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquiring lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.706918] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Acquired lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.707168] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.720307] env[61947]: ERROR nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 711.720307] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.720307] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.720307] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.720307] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.720307] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.720307] env[61947]: ERROR nova.compute.manager raise self.value [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.720307] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.720307] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.720307] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.720749] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.720749] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.720749] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 711.720749] env[61947]: ERROR nova.compute.manager [ 711.720749] env[61947]: Traceback (most recent call last): [ 711.720749] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.720749] env[61947]: listener.cb(fileno) [ 711.720749] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.720749] env[61947]: result = function(*args, **kwargs) [ 711.720749] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.720749] env[61947]: return func(*args, **kwargs) [ 711.720749] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.720749] env[61947]: raise e [ 711.720749] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.720749] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 711.720749] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.720749] env[61947]: created_port_ids = self._update_ports_for_instance( [ 711.720749] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.720749] env[61947]: with excutils.save_and_reraise_exception(): [ 711.720749] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.720749] env[61947]: self.force_reraise() [ 711.720749] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.720749] env[61947]: raise self.value [ 711.720749] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.720749] env[61947]: updated_port = self._update_port( [ 711.720749] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.720749] env[61947]: _ensure_no_port_binding_failure(port) [ 711.720749] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.720749] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.721564] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 711.721564] env[61947]: Removing descriptor: 17 [ 711.763884] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.790834] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.791019] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.791192] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.791386] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.791530] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.791676] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.791880] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.792159] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.792364] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.792594] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.793729] env[61947]: DEBUG nova.virt.hardware [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.794691] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174640fb-2384-460f-9173-d73a5d74b190 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.806363] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bee6e8-9aa2-4515-82ea-8097b701eed3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.821377] env[61947]: ERROR nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Traceback (most recent call last): [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] yield resources [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.driver.spawn(context, instance, image_meta, [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] vm_ref = self.build_virtual_machine(instance, [ 711.821377] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] for vif in network_info: [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return self._sync_wrapper(fn, *args, **kwargs) [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.wait() [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self[:] = self._gt.wait() [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return self._exit_event.wait() [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.821690] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] current.throw(*self._exc) [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] result = function(*args, **kwargs) [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return func(*args, **kwargs) [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise e [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] nwinfo = self.network_api.allocate_for_instance( [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] created_port_ids = self._update_ports_for_instance( [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] with excutils.save_and_reraise_exception(): [ 711.822152] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.force_reraise() [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise self.value [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] updated_port = self._update_port( [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] _ensure_no_port_binding_failure(port) [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise exception.PortBindingFailed(port_id=port['id']) [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 711.822487] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] [ 711.822487] env[61947]: INFO nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Terminating instance [ 711.823570] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquiring lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.985119] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.062345] env[61947]: DEBUG nova.network.neutron [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.150373] env[61947]: DEBUG nova.network.neutron [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.165861] env[61947]: INFO nova.compute.manager [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Took 39.87 seconds to build instance. [ 712.227023] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.351121] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.584231] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a93274b-eb48-46a1-ad95-acb14f0c183d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.592680] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0aa86b-527f-42cb-99e0-900a7e3943af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.627026] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51af66d-95fd-4a6c-86ed-2c45a489292c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.634728] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfe3523-e688-4368-a7f8-857ed9af201e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.647768] env[61947]: DEBUG nova.compute.provider_tree [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.655506] env[61947]: DEBUG oslo_concurrency.lockutils [req-a0123b92-df92-4810-9c61-1453646dc26d req-bfc96acd-d8d8-439f-8c79-b8b30b2699df service nova] Releasing lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.655506] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquired lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.655506] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.669162] env[61947]: DEBUG oslo_concurrency.lockutils [None req-50798c93-b8ab-46db-83e8-eb1a5782c346 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.039s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.856224] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Releasing lock "refresh_cache-b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.856480] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.856669] env[61947]: DEBUG nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.856837] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.872681] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.150878] env[61947]: DEBUG nova.scheduler.client.report [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.171426] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.175569] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.244852] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.375141] env[61947]: DEBUG nova.network.neutron [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.434726] env[61947]: INFO nova.compute.manager [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Rebuilding instance [ 713.477602] env[61947]: DEBUG nova.compute.manager [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 713.478501] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45f6531-67de-4777-b1c9-a28f81e16b8a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.565687] env[61947]: DEBUG nova.compute.manager [req-9944d62e-09db-4bd3-8519-a3ca3bd712f1 req-bf87a9ee-d776-4aa3-beaa-9d59139d8005 service nova] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Received event network-vif-deleted-34ca4adb-5227-4735-99d9-7d7293987aec {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.659676] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.660335] env[61947]: ERROR nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Traceback (most recent call last): [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.driver.spawn(context, instance, image_meta, [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] vm_ref = self.build_virtual_machine(instance, [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.660335] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] for vif in network_info: [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self._sync_wrapper(fn, *args, **kwargs) [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.wait() [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self[:] = self._gt.wait() [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self._exit_event.wait() [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] result = hub.switch() [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.660647] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return self.greenlet.switch() [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] result = function(*args, **kwargs) [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] return func(*args, **kwargs) [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise e [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] nwinfo = self.network_api.allocate_for_instance( [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] created_port_ids = self._update_ports_for_instance( [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] with excutils.save_and_reraise_exception(): [ 713.661041] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] self.force_reraise() [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise self.value [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] updated_port = self._update_port( [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] _ensure_no_port_binding_failure(port) [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] raise exception.PortBindingFailed(port_id=port['id']) [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] nova.exception.PortBindingFailed: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. [ 713.661348] env[61947]: ERROR nova.compute.manager [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] [ 713.661609] env[61947]: DEBUG nova.compute.utils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.662362] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.563s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.665214] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Build of instance 6ca2d77f-a62e-4eff-bf15-7855c94bb649 was re-scheduled: Binding failed for port ac3204b8-5723-41cf-9118-2fc08248828c, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.665659] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.665884] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.666045] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.666207] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.697094] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.748953] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Releasing lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.748953] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 713.748953] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 713.749316] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ceccdd2d-2a41-4d02-a79c-cd7046a1d969 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.762899] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfad3334-93a4-4159-b75e-85663655a06f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.786479] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d58f3c69-ae7a-48d5-a796-34898c6e704c could not be found. [ 713.786715] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 713.786895] env[61947]: INFO nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 713.787153] env[61947]: DEBUG oslo.service.loopingcall [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.787373] env[61947]: DEBUG nova.compute.manager [-] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.787464] env[61947]: DEBUG nova.network.neutron [-] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.817984] env[61947]: DEBUG nova.network.neutron [-] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.878761] env[61947]: INFO nova.compute.manager [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] [instance: b33f5bbe-fea2-40e1-bdcc-5c4788568e7d] Took 1.02 seconds to deallocate network for instance. [ 713.991927] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 713.991927] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2e3f4ef-332b-49d0-b215-d9b383267e47 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.999117] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 713.999117] env[61947]: value = "task-1224178" [ 713.999117] env[61947]: _type = "Task" [ 713.999117] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.009218] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.200013] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.322982] env[61947]: DEBUG nova.network.neutron [-] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.353735] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.511510] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224178, 'name': PowerOffVM_Task, 'duration_secs': 0.13538} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.511510] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 714.511729] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 714.512511] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8bfd7d-f79d-417a-a8d4-f366cc118ae0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.519147] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 714.521544] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4667258b-799f-490b-8c51-735c1c410ed0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.543637] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9adaf9-2ecf-48a6-bab1-fd8a26e62cda {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.547607] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 714.547910] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 714.548275] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleting the datastore file [datastore2] 32fe176d-ebc2-46a4-8e93-3b2555a681d4 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 714.549597] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0011aee0-83f4-4493-bd23-f4cd1b03a4f8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.553996] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef0b1b6-5edb-48ac-b234-5f6655797783 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.558356] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 714.558356] env[61947]: value = "task-1224180" [ 714.558356] env[61947]: _type = "Task" [ 714.558356] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.590969] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87eba0b7-f36c-4628-9bed-d343d32e8b52 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.593462] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.598354] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a179f597-5435-484c-9807-c39725a00de5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.611336] env[61947]: DEBUG nova.compute.provider_tree [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.826478] env[61947]: INFO nova.compute.manager [-] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Took 1.04 seconds to deallocate network for instance. [ 714.829327] env[61947]: DEBUG nova.compute.claims [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 714.829512] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.856343] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "refresh_cache-6ca2d77f-a62e-4eff-bf15-7855c94bb649" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.856590] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.856775] env[61947]: DEBUG nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.856948] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.883137] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.908821] env[61947]: INFO nova.scheduler.client.report [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Deleted allocations for instance b33f5bbe-fea2-40e1-bdcc-5c4788568e7d [ 715.069074] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102158} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.069362] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 715.069743] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 715.069939] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 715.114550] env[61947]: DEBUG nova.scheduler.client.report [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.389238] env[61947]: DEBUG nova.network.neutron [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.419058] env[61947]: DEBUG oslo_concurrency.lockutils [None req-576cf42d-52b7-4f19-96ba-6b21da290512 tempest-ServersTestFqdnHostnames-2066274474 tempest-ServersTestFqdnHostnames-2066274474-project-member] Lock "b33f5bbe-fea2-40e1-bdcc-5c4788568e7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.680s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.619480] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.620180] env[61947]: ERROR nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] Traceback (most recent call last): [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.driver.spawn(context, instance, image_meta, [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] vm_ref = self.build_virtual_machine(instance, [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.620180] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] for vif in network_info: [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self._sync_wrapper(fn, *args, **kwargs) [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.wait() [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self[:] = self._gt.wait() [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self._exit_event.wait() [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] result = hub.switch() [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 715.620852] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return self.greenlet.switch() [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] result = function(*args, **kwargs) [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] return func(*args, **kwargs) [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise e [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] nwinfo = self.network_api.allocate_for_instance( [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] created_port_ids = self._update_ports_for_instance( [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] with excutils.save_and_reraise_exception(): [ 715.621395] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] self.force_reraise() [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise self.value [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] updated_port = self._update_port( [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] _ensure_no_port_binding_failure(port) [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] raise exception.PortBindingFailed(port_id=port['id']) [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] nova.exception.PortBindingFailed: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. [ 715.622273] env[61947]: ERROR nova.compute.manager [instance: 2eaab038-6c55-439a-8293-da80474d369a] [ 715.622872] env[61947]: DEBUG nova.compute.utils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.622872] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.123s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.623502] env[61947]: INFO nova.compute.claims [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.626590] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Build of instance 2eaab038-6c55-439a-8293-da80474d369a was re-scheduled: Binding failed for port 5b6f6ca7-bf10-4fc4-a69e-b1c85980df1d, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.627022] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.627246] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.627389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.627545] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.892395] env[61947]: INFO nova.compute.manager [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: 6ca2d77f-a62e-4eff-bf15-7855c94bb649] Took 1.03 seconds to deallocate network for instance. [ 715.922364] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.104429] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.104631] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.104768] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.104958] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.105162] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.105323] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.105532] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.105695] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.105945] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.106025] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.106197] env[61947]: DEBUG nova.virt.hardware [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.107107] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d789fd8-a4e4-4992-9b35-4cf1551a5bd7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.115348] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e020bf-62ae-4f3b-8c2e-3ac0fbdda148 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.131905] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.137103] env[61947]: DEBUG oslo.service.loopingcall [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.139778] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 716.140591] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3345acb-b133-4193-bebc-8e829ded988d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.158743] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.158743] env[61947]: value = "task-1224181" [ 716.158743] env[61947]: _type = "Task" [ 716.158743] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.166349] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.171385] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224181, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.266573] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.450386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.669163] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224181, 'name': CreateVM_Task, 'duration_secs': 0.271899} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.671865] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.672952] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.672952] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.673576] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.673911] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-420eaa37-5479-4921-81a3-ad8125d88256 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.679250] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 716.679250] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b26bbf-150c-aa1e-da85-948110b5d19d" [ 716.679250] env[61947]: _type = "Task" [ 716.679250] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.690638] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b26bbf-150c-aa1e-da85-948110b5d19d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.757424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.758069] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.773493] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-2eaab038-6c55-439a-8293-da80474d369a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.773716] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.773894] env[61947]: DEBUG nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.774071] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.794638] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.932330] env[61947]: INFO nova.scheduler.client.report [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted allocations for instance 6ca2d77f-a62e-4eff-bf15-7855c94bb649 [ 717.049947] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77084cd-9225-4925-8a76-28ca12e7d2fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.058940] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f139d5-e855-40a0-b465-7c6b90bb8bc4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.092015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613093b8-4865-4a7c-b8ff-d992e5e16678 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.098192] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde83950-23ab-4aee-b63d-a8357af0b3d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.114049] env[61947]: DEBUG nova.compute.provider_tree [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.190249] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b26bbf-150c-aa1e-da85-948110b5d19d, 'name': SearchDatastore_Task, 'duration_secs': 0.009582} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.190540] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.190770] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.191010] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.191163] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.191482] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.191579] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5da70372-d09c-48eb-b342-73280b81e511 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.199213] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.199381] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 717.200058] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a2adcab-0945-464d-975b-a7eb6922eee5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.204737] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 717.204737] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528da6f2-171a-0862-6c7c-d045dcf8a455" [ 717.204737] env[61947]: _type = "Task" [ 717.204737] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.212552] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528da6f2-171a-0862-6c7c-d045dcf8a455, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.298351] env[61947]: DEBUG nova.network.neutron [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.445932] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36273c39-012f-4d12-a7dd-16a8344543f6 tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "6ca2d77f-a62e-4eff-bf15-7855c94bb649" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.397s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.616539] env[61947]: DEBUG nova.scheduler.client.report [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.717036] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528da6f2-171a-0862-6c7c-d045dcf8a455, 'name': SearchDatastore_Task, 'duration_secs': 0.008279} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.717894] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa248991-eda8-41c2-b92a-841a4e2d7364 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.726637] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 717.726637] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ef28d5-2e2e-5d3b-d5f2-830e5e3999c5" [ 717.726637] env[61947]: _type = "Task" [ 717.726637] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.732305] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ef28d5-2e2e-5d3b-d5f2-830e5e3999c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.801168] env[61947]: INFO nova.compute.manager [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 2eaab038-6c55-439a-8293-da80474d369a] Took 1.03 seconds to deallocate network for instance. [ 717.951682] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.123166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.123626] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.126318] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.729s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.128066] env[61947]: INFO nova.compute.claims [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.240991] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ef28d5-2e2e-5d3b-d5f2-830e5e3999c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009107} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.241393] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.241728] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 718.242084] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54c27024-25e6-4424-b636-29367e6033d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.250049] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 718.250049] env[61947]: value = "task-1224182" [ 718.250049] env[61947]: _type = "Task" [ 718.250049] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.261308] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224182, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.481210] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.633385] env[61947]: DEBUG nova.compute.utils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.637155] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 718.637909] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.716956] env[61947]: DEBUG nova.policy [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10d22f3901274e16a22df21876a65820', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5f4616d734c452b851cae554c6ec4ef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 718.763512] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224182, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46704} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.763772] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 718.763981] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.764248] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b94fa580-43ac-4aed-87dc-5e21f609307c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.770351] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 718.770351] env[61947]: value = "task-1224183" [ 718.770351] env[61947]: _type = "Task" [ 718.770351] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.778448] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.834737] env[61947]: INFO nova.scheduler.client.report [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance 2eaab038-6c55-439a-8293-da80474d369a [ 719.140226] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.280764] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060038} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.281051] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.281830] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22eb1842-0684-413d-a438-15173da435eb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.303162] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.306523] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f91c7423-f6b4-41ad-9ce8-aabeeaae30ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.322932] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Successfully created port: a07c729e-d13d-46ab-b727-03b40b2b586c {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.332675] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 719.332675] env[61947]: value = "task-1224184" [ 719.332675] env[61947]: _type = "Task" [ 719.332675] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.345324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a971fd23-6b33-4c65-afd9-2947216db6e8 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "2eaab038-6c55-439a-8293-da80474d369a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.670s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.345581] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224184, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.572375] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18625e5-1bf3-45d0-91f5-271e0a28763d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.580317] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617f577a-e340-4372-9fb6-418702ef8cd5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.619076] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7d6eb2-17c5-41c3-ab55-e8a4ec956ef0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.626780] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf4816e-3d96-433b-a13e-f5149f99ff80 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.642938] env[61947]: DEBUG nova.compute.provider_tree [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.846033] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224184, 'name': ReconfigVM_Task, 'duration_secs': 0.273872} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.846033] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 32fe176d-ebc2-46a4-8e93-3b2555a681d4/32fe176d-ebc2-46a4-8e93-3b2555a681d4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.846033] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68811920-cb31-44f6-b2bd-37111b10d5dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.847880] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.856403] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 719.856403] env[61947]: value = "task-1224185" [ 719.856403] env[61947]: _type = "Task" [ 719.856403] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.866389] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224185, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.147444] env[61947]: DEBUG nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.154133] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.187087] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.187341] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.187687] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.187942] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.188156] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.188349] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.188700] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.188919] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.189127] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.189471] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.189766] env[61947]: DEBUG nova.virt.hardware [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.190637] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c968be4a-19c5-4e4b-980b-72c16282bafa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.201895] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d1fce8-5f77-45f7-b4d7-bb288742b03a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.367646] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224185, 'name': Rename_Task, 'duration_secs': 0.136907} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.368505] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.368801] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.369062] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d773cc5d-73ba-43b7-ad92-cd2c85ac6610 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.375202] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 720.375202] env[61947]: value = "task-1224186" [ 720.375202] env[61947]: _type = "Task" [ 720.375202] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.383038] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.470091] env[61947]: DEBUG nova.compute.manager [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Received event network-changed-a07c729e-d13d-46ab-b727-03b40b2b586c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 720.470308] env[61947]: DEBUG nova.compute.manager [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Refreshing instance network info cache due to event network-changed-a07c729e-d13d-46ab-b727-03b40b2b586c. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 720.470524] env[61947]: DEBUG oslo_concurrency.lockutils [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] Acquiring lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.470665] env[61947]: DEBUG oslo_concurrency.lockutils [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] Acquired lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.470856] env[61947]: DEBUG nova.network.neutron [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Refreshing network info cache for port a07c729e-d13d-46ab-b727-03b40b2b586c {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.476903] env[61947]: ERROR nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 720.476903] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.476903] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.476903] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.476903] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.476903] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.476903] env[61947]: ERROR nova.compute.manager raise self.value [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.476903] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.476903] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.476903] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.477399] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.477399] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.477399] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 720.477399] env[61947]: ERROR nova.compute.manager [ 720.477399] env[61947]: Traceback (most recent call last): [ 720.477399] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.477399] env[61947]: listener.cb(fileno) [ 720.477399] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.477399] env[61947]: result = function(*args, **kwargs) [ 720.477399] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.477399] env[61947]: return func(*args, **kwargs) [ 720.477399] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.477399] env[61947]: raise e [ 720.477399] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.477399] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 720.477399] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.477399] env[61947]: created_port_ids = self._update_ports_for_instance( [ 720.477399] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.477399] env[61947]: with excutils.save_and_reraise_exception(): [ 720.477399] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.477399] env[61947]: self.force_reraise() [ 720.477399] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.477399] env[61947]: raise self.value [ 720.477399] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.477399] env[61947]: updated_port = self._update_port( [ 720.477399] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.477399] env[61947]: _ensure_no_port_binding_failure(port) [ 720.477399] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.477399] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.478373] env[61947]: nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 720.478373] env[61947]: Removing descriptor: 17 [ 720.478373] env[61947]: ERROR nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] Traceback (most recent call last): [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] yield resources [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.driver.spawn(context, instance, image_meta, [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.478373] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] vm_ref = self.build_virtual_machine(instance, [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] for vif in network_info: [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self._sync_wrapper(fn, *args, **kwargs) [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.wait() [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self[:] = self._gt.wait() [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self._exit_event.wait() [ 720.478719] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] result = hub.switch() [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self.greenlet.switch() [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] result = function(*args, **kwargs) [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return func(*args, **kwargs) [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise e [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] nwinfo = self.network_api.allocate_for_instance( [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.479156] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] created_port_ids = self._update_ports_for_instance( [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] with excutils.save_and_reraise_exception(): [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.force_reraise() [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise self.value [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] updated_port = self._update_port( [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] _ensure_no_port_binding_failure(port) [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.479519] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise exception.PortBindingFailed(port_id=port['id']) [ 720.479854] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 720.479854] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] [ 720.479854] env[61947]: INFO nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Terminating instance [ 720.479854] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquiring lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.652894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.655234] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.656211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.463s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.658045] env[61947]: INFO nova.compute.claims [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.762188] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.762699] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.888522] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224186, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.999135] env[61947]: DEBUG nova.network.neutron [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.123242] env[61947]: DEBUG nova.network.neutron [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.163412] env[61947]: DEBUG nova.compute.utils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.171279] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.171279] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.218804] env[61947]: DEBUG nova.policy [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05b9a0659cf14e4883a3400d6da44fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d53677e3f024ef38a7868b1d2a82bc4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 721.390352] env[61947]: DEBUG oslo_vmware.api [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224186, 'name': PowerOnVM_Task, 'duration_secs': 0.691689} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.390671] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.390876] env[61947]: DEBUG nova.compute.manager [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.391699] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcbbbc8-25d9-4f9d-8510-cdb5a52ec5f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.553413] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Successfully created port: c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.627845] env[61947]: DEBUG oslo_concurrency.lockutils [req-44ea198f-7fd3-4420-8492-e008b71d349b req-e330699b-ffd5-4efb-beb7-7e83e82cde7a service nova] Releasing lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.628365] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquired lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.628517] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.668855] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.882957] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.883037] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.914235] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.133116] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ff9e47-f9f5-40ee-8f7e-1e2e82ead408 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.141221] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6403e712-add4-484e-a847-0525e2120b0e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.172450] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.174775] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b39b7dc-c5b2-4e8a-998b-adfb9f251ecc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.185563] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aec70da-a7ee-4ff1-a6fc-3c3639fcfe98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.199519] env[61947]: DEBUG nova.compute.provider_tree [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.317234] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.590139] env[61947]: DEBUG nova.compute.manager [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: 947b697d-4120-491d-83ce-21681974c224] Received event network-vif-deleted-a07c729e-d13d-46ab-b727-03b40b2b586c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.590286] env[61947]: DEBUG nova.compute.manager [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Received event network-changed-c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.590455] env[61947]: DEBUG nova.compute.manager [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Refreshing instance network info cache due to event network-changed-c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 722.590670] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] Acquiring lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.590809] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] Acquired lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.590965] env[61947]: DEBUG nova.network.neutron [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Refreshing network info cache for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.690759] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.693982] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 722.693982] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.693982] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.693982] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.693982] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.693982] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.693982] env[61947]: ERROR nova.compute.manager raise self.value [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.693982] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.693982] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.693982] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.695276] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.695276] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.695276] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 722.695276] env[61947]: ERROR nova.compute.manager [ 722.695276] env[61947]: Traceback (most recent call last): [ 722.695276] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.695276] env[61947]: listener.cb(fileno) [ 722.695276] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.695276] env[61947]: result = function(*args, **kwargs) [ 722.695276] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.695276] env[61947]: return func(*args, **kwargs) [ 722.695276] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.695276] env[61947]: raise e [ 722.695276] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.695276] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 722.695276] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.695276] env[61947]: created_port_ids = self._update_ports_for_instance( [ 722.695276] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.695276] env[61947]: with excutils.save_and_reraise_exception(): [ 722.695276] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.695276] env[61947]: self.force_reraise() [ 722.695276] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.695276] env[61947]: raise self.value [ 722.695276] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.695276] env[61947]: updated_port = self._update_port( [ 722.695276] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.695276] env[61947]: _ensure_no_port_binding_failure(port) [ 722.695276] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.695276] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.696196] env[61947]: nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 722.696196] env[61947]: Removing descriptor: 17 [ 722.702193] env[61947]: DEBUG nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.714767] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.715044] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.715223] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.715407] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.715550] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.715751] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.716009] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.716207] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.716447] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.716539] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.716709] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.717590] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7850a7c-c694-45a6-bd17-1fe29c8af605 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.726703] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3db5949-0980-4f1a-9210-9e29b6608d05 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.741180] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Traceback (most recent call last): [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] yield resources [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.driver.spawn(context, instance, image_meta, [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] vm_ref = self.build_virtual_machine(instance, [ 722.741180] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] for vif in network_info: [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return self._sync_wrapper(fn, *args, **kwargs) [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.wait() [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self[:] = self._gt.wait() [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return self._exit_event.wait() [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.741555] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] current.throw(*self._exc) [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] result = function(*args, **kwargs) [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return func(*args, **kwargs) [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise e [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] nwinfo = self.network_api.allocate_for_instance( [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] created_port_ids = self._update_ports_for_instance( [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] with excutils.save_and_reraise_exception(): [ 722.741931] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.force_reraise() [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise self.value [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] updated_port = self._update_port( [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] _ensure_no_port_binding_failure(port) [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise exception.PortBindingFailed(port_id=port['id']) [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 722.742324] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] [ 722.742324] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Terminating instance [ 722.745744] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.824121] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Releasing lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.824121] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.824121] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 722.824121] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6672f6a5-b680-4207-94ea-cecebefec1db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.831182] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94921c5a-91f4-4660-b7e8-b733c12ef699 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.841832] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.842204] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.842653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.842956] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.843247] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.845495] env[61947]: INFO nova.compute.manager [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Terminating instance [ 722.851028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "refresh_cache-32fe176d-ebc2-46a4-8e93-3b2555a681d4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.851305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "refresh_cache-32fe176d-ebc2-46a4-8e93-3b2555a681d4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.851579] env[61947]: DEBUG nova.network.neutron [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.858259] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 947b697d-4120-491d-83ce-21681974c224 could not be found. [ 722.858571] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 722.858855] env[61947]: INFO nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.859204] env[61947]: DEBUG oslo.service.loopingcall [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.860015] env[61947]: DEBUG nova.compute.manager [-] [instance: 947b697d-4120-491d-83ce-21681974c224] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.860219] env[61947]: DEBUG nova.network.neutron [-] [instance: 947b697d-4120-491d-83ce-21681974c224] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.041463] env[61947]: DEBUG nova.network.neutron [-] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.116167] env[61947]: DEBUG nova.network.neutron [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.206957] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.207514] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.210725] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.956s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.212371] env[61947]: INFO nova.compute.claims [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.230481] env[61947]: DEBUG nova.network.neutron [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.377937] env[61947]: DEBUG nova.network.neutron [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.435569] env[61947]: DEBUG nova.network.neutron [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.542986] env[61947]: DEBUG nova.network.neutron [-] [instance: 947b697d-4120-491d-83ce-21681974c224] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.718787] env[61947]: DEBUG nova.compute.utils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.725730] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.725730] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.734309] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9fd9a98-79a9-411d-8e66-302520a2f7b5 req-c0bfafb3-2279-4559-b1a8-866d691e7a58 service nova] Releasing lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.734473] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.734636] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.767732] env[61947]: DEBUG nova.policy [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05b9a0659cf14e4883a3400d6da44fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d53677e3f024ef38a7868b1d2a82bc4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 723.940348] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "refresh_cache-32fe176d-ebc2-46a4-8e93-3b2555a681d4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.940903] env[61947]: DEBUG nova.compute.manager [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.941162] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.941531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.942402] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.943255] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26caca43-5a9f-4191-9f6e-caf23e57ad6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.952561] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 723.953588] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6bb07a51-dee3-4039-b917-0191feac90e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.961091] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 723.961091] env[61947]: value = "task-1224187" [ 723.961091] env[61947]: _type = "Task" [ 723.961091] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.970750] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224187, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.046915] env[61947]: INFO nova.compute.manager [-] [instance: 947b697d-4120-491d-83ce-21681974c224] Took 1.19 seconds to deallocate network for instance. [ 724.053023] env[61947]: DEBUG nova.compute.claims [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 724.053023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.090236] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Successfully created port: da42e119-9f9c-4444-8733-fef0693a8338 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.226715] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.261963] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.402187] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.473241] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224187, 'name': PowerOffVM_Task, 'duration_secs': 0.185398} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.473241] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.473241] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.473241] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74ef717c-6850-4f5e-94a0-406c690122f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.498453] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.498669] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.498877] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleting the datastore file [datastore1] 32fe176d-ebc2-46a4-8e93-3b2555a681d4 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.499147] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c37adff-58b2-4c15-8ed9-b7c9049cc422 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.510835] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 724.510835] env[61947]: value = "task-1224189" [ 724.510835] env[61947]: _type = "Task" [ 724.510835] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.523242] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.669761] env[61947]: DEBUG nova.compute.manager [req-52c43121-0869-4ecf-802e-337df6bd6ad8 req-15de0a28-1969-418d-a92e-3a7c7b116d26 service nova] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Received event network-vif-deleted-c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.682385] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89dc0e2e-79c0-4f27-bfbf-14a588bab0cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.688339] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033aa751-f3af-4699-bb6a-128a88123ba9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.721481] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3391c21-272c-4bc5-a7d2-30aef2ba314b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.729343] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee35cd9d-2ab2-4d24-a93b-368baafd1af0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.746085] env[61947]: DEBUG nova.compute.provider_tree [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.904833] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.908018] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.908018] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 724.908018] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc720a4e-5920-4026-9d5a-6ed018b9f3e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.914427] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b780dc4f-0339-4e90-bfb7-a6b01e270125 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.936859] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8c66631-1fae-4a03-810f-66d64a75b118 could not be found. [ 724.937131] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.937457] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Took 0.03 seconds to destroy the instance on the hypervisor. [ 724.937552] env[61947]: DEBUG oslo.service.loopingcall [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.937786] env[61947]: DEBUG nova.compute.manager [-] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.937890] env[61947]: DEBUG nova.network.neutron [-] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.956058] env[61947]: DEBUG nova.network.neutron [-] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.024103] env[61947]: DEBUG oslo_vmware.api [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095776} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.024103] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.024103] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 725.024103] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.024103] env[61947]: INFO nova.compute.manager [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Took 1.08 seconds to destroy the instance on the hypervisor. [ 725.024442] env[61947]: DEBUG oslo.service.loopingcall [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.024442] env[61947]: DEBUG nova.compute.manager [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.024442] env[61947]: DEBUG nova.network.neutron [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.038934] env[61947]: DEBUG nova.network.neutron [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.183039] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 725.183039] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.183039] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.183039] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.183039] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.183039] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.183039] env[61947]: ERROR nova.compute.manager raise self.value [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.183039] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.183039] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.183039] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.183541] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.183541] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.183541] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 725.183541] env[61947]: ERROR nova.compute.manager [ 725.183541] env[61947]: Traceback (most recent call last): [ 725.183541] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.183541] env[61947]: listener.cb(fileno) [ 725.183541] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.183541] env[61947]: result = function(*args, **kwargs) [ 725.183541] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.183541] env[61947]: return func(*args, **kwargs) [ 725.183541] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.183541] env[61947]: raise e [ 725.183541] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.183541] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 725.183541] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.183541] env[61947]: created_port_ids = self._update_ports_for_instance( [ 725.183541] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.183541] env[61947]: with excutils.save_and_reraise_exception(): [ 725.183541] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.183541] env[61947]: self.force_reraise() [ 725.183541] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.183541] env[61947]: raise self.value [ 725.183541] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.183541] env[61947]: updated_port = self._update_port( [ 725.183541] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.183541] env[61947]: _ensure_no_port_binding_failure(port) [ 725.183541] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.183541] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.184373] env[61947]: nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 725.184373] env[61947]: Removing descriptor: 17 [ 725.249091] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.251670] env[61947]: DEBUG nova.scheduler.client.report [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.280286] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.280553] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.280708] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.280887] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.281043] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.281192] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.281397] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.281553] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.281724] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.281884] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.282064] env[61947]: DEBUG nova.virt.hardware [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.283119] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14044e7-f5d9-47cf-8eb6-56cba514846a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.291293] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d783b6fd-fd50-48c8-95ca-5c8506892947 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.306225] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Traceback (most recent call last): [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] yield resources [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.driver.spawn(context, instance, image_meta, [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] vm_ref = self.build_virtual_machine(instance, [ 725.306225] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] for vif in network_info: [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return self._sync_wrapper(fn, *args, **kwargs) [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.wait() [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self[:] = self._gt.wait() [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return self._exit_event.wait() [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.306577] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] current.throw(*self._exc) [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] result = function(*args, **kwargs) [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return func(*args, **kwargs) [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise e [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] nwinfo = self.network_api.allocate_for_instance( [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] created_port_ids = self._update_ports_for_instance( [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] with excutils.save_and_reraise_exception(): [ 725.306952] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.force_reraise() [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise self.value [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] updated_port = self._update_port( [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] _ensure_no_port_binding_failure(port) [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise exception.PortBindingFailed(port_id=port['id']) [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 725.307338] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] [ 725.307338] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Terminating instance [ 725.308405] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.308586] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.308757] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.458678] env[61947]: DEBUG nova.network.neutron [-] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.543377] env[61947]: DEBUG nova.network.neutron [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.756051] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.756601] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.759597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.774s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.760739] env[61947]: INFO nova.compute.claims [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.825302] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.918109] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.961880] env[61947]: INFO nova.compute.manager [-] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Took 1.02 seconds to deallocate network for instance. [ 725.965979] env[61947]: DEBUG nova.compute.claims [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 725.966250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.045294] env[61947]: INFO nova.compute.manager [-] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Took 1.02 seconds to deallocate network for instance. [ 726.265953] env[61947]: DEBUG nova.compute.utils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.269511] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.269675] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.309585] env[61947]: DEBUG nova.policy [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 726.421123] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.421565] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.421783] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 726.422100] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc3f2950-82ae-44c1-a5bc-27d98f02ed96 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.433055] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42a32d2-0444-4b19-afd4-6545c67303e0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.455382] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37 could not be found. [ 726.455515] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.455691] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Took 0.03 seconds to destroy the instance on the hypervisor. [ 726.455843] env[61947]: DEBUG oslo.service.loopingcall [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.456084] env[61947]: DEBUG nova.compute.manager [-] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.456849] env[61947]: DEBUG nova.network.neutron [-] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.479100] env[61947]: DEBUG nova.network.neutron [-] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.552688] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.661914] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Successfully created port: bf2d79c6-1c15-48ac-a876-92e0169f8fa2 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.716251] env[61947]: DEBUG nova.compute.manager [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Received event network-changed-da42e119-9f9c-4444-8733-fef0693a8338 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.716485] env[61947]: DEBUG nova.compute.manager [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Refreshing instance network info cache due to event network-changed-da42e119-9f9c-4444-8733-fef0693a8338. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 726.716703] env[61947]: DEBUG oslo_concurrency.lockutils [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] Acquiring lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.716849] env[61947]: DEBUG oslo_concurrency.lockutils [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] Acquired lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.717023] env[61947]: DEBUG nova.network.neutron [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Refreshing network info cache for port da42e119-9f9c-4444-8733-fef0693a8338 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.772637] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.985354] env[61947]: DEBUG nova.network.neutron [-] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.142075] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdfcf29-2a89-4bd3-b88f-a46b8601bea4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.150019] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7feaf70f-64a7-4546-8467-1d98566868ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.182852] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259c62e9-ec51-4eb9-9b85-36ab84ff8e75 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.190347] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad17988e-e769-4c00-9452-438f624024ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.203741] env[61947]: DEBUG nova.compute.provider_tree [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.236676] env[61947]: DEBUG nova.network.neutron [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.344762] env[61947]: DEBUG nova.network.neutron [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.487988] env[61947]: INFO nova.compute.manager [-] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Took 1.03 seconds to deallocate network for instance. [ 727.492674] env[61947]: DEBUG nova.compute.claims [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 727.492878] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.691256] env[61947]: ERROR nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 727.691256] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.691256] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.691256] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.691256] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.691256] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.691256] env[61947]: ERROR nova.compute.manager raise self.value [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.691256] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.691256] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.691256] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.691720] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.691720] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.691720] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 727.691720] env[61947]: ERROR nova.compute.manager [ 727.691720] env[61947]: Traceback (most recent call last): [ 727.691720] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.691720] env[61947]: listener.cb(fileno) [ 727.691720] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.691720] env[61947]: result = function(*args, **kwargs) [ 727.691720] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.691720] env[61947]: return func(*args, **kwargs) [ 727.691720] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.691720] env[61947]: raise e [ 727.691720] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.691720] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 727.691720] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.691720] env[61947]: created_port_ids = self._update_ports_for_instance( [ 727.691720] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.691720] env[61947]: with excutils.save_and_reraise_exception(): [ 727.691720] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.691720] env[61947]: self.force_reraise() [ 727.691720] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.691720] env[61947]: raise self.value [ 727.691720] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.691720] env[61947]: updated_port = self._update_port( [ 727.691720] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.691720] env[61947]: _ensure_no_port_binding_failure(port) [ 727.691720] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.691720] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.692667] env[61947]: nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 727.692667] env[61947]: Removing descriptor: 17 [ 727.706637] env[61947]: DEBUG nova.scheduler.client.report [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.787033] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.813958] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.814232] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.814388] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.814566] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.814709] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.814853] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.815077] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.815241] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.815407] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.815564] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.815736] env[61947]: DEBUG nova.virt.hardware [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.816624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bd95e6-74cc-4aa1-a113-73e1b9472949 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.824691] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05079fc4-ffad-40f5-aa21-a958deac53e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.840371] env[61947]: ERROR nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Traceback (most recent call last): [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] yield resources [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.driver.spawn(context, instance, image_meta, [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] vm_ref = self.build_virtual_machine(instance, [ 727.840371] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] for vif in network_info: [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return self._sync_wrapper(fn, *args, **kwargs) [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.wait() [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self[:] = self._gt.wait() [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return self._exit_event.wait() [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 727.840716] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] current.throw(*self._exc) [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] result = function(*args, **kwargs) [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return func(*args, **kwargs) [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise e [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] nwinfo = self.network_api.allocate_for_instance( [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] created_port_ids = self._update_ports_for_instance( [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] with excutils.save_and_reraise_exception(): [ 727.841044] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.force_reraise() [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise self.value [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] updated_port = self._update_port( [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] _ensure_no_port_binding_failure(port) [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise exception.PortBindingFailed(port_id=port['id']) [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 727.841439] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] [ 727.841439] env[61947]: INFO nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Terminating instance [ 727.842806] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.842966] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.843206] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.847034] env[61947]: DEBUG oslo_concurrency.lockutils [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] Releasing lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.847034] env[61947]: DEBUG nova.compute.manager [req-16c2ac68-5749-4263-b7d0-56461b9ba915 req-e817c8f1-7a2c-4004-bf85-142817e7d2c9 service nova] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Received event network-vif-deleted-da42e119-9f9c-4444-8733-fef0693a8338 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.212160] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.212692] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.215712] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.519s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.217224] env[61947]: INFO nova.compute.claims [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.362877] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.444999] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.721832] env[61947]: DEBUG nova.compute.utils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.725806] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.726115] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.737160] env[61947]: DEBUG nova.compute.manager [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Received event network-changed-bf2d79c6-1c15-48ac-a876-92e0169f8fa2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.737308] env[61947]: DEBUG nova.compute.manager [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Refreshing instance network info cache due to event network-changed-bf2d79c6-1c15-48ac-a876-92e0169f8fa2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 728.737498] env[61947]: DEBUG oslo_concurrency.lockutils [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] Acquiring lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.764616] env[61947]: DEBUG nova.policy [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0337baafe604194a5bc93e99ca8078a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51aa57efe97e453783044286f33d3f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 728.948348] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.948787] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.949014] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 728.949343] env[61947]: DEBUG oslo_concurrency.lockutils [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] Acquired lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.949514] env[61947]: DEBUG nova.network.neutron [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Refreshing network info cache for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.950979] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26a2b0e4-6fbd-431e-a803-be9f7e2cc740 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.959707] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39318845-5483-4484-a08f-57c3f2c396f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.982962] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a4df3f1-67f3-404d-953f-caa6ba1b1641 could not be found. [ 728.983268] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 728.983431] env[61947]: INFO nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.983696] env[61947]: DEBUG oslo.service.loopingcall [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.983936] env[61947]: DEBUG nova.compute.manager [-] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.984058] env[61947]: DEBUG nova.network.neutron [-] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.004833] env[61947]: DEBUG nova.network.neutron [-] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.056153] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Successfully created port: 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.228793] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.477144] env[61947]: DEBUG nova.network.neutron [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.507453] env[61947]: DEBUG nova.network.neutron [-] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.577085] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c2b209-c740-4187-bb7c-141b8c86ee3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.581816] env[61947]: DEBUG nova.network.neutron [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.586632] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cca6a2-425e-43a5-8aa8-c93d3f424296 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.616806] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cf6433-5502-4c27-b8a3-ee60b037c94e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.626620] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdd3e40-5ee7-4449-b0ae-d4b0fbab43c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.646533] env[61947]: DEBUG nova.compute.provider_tree [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.010192] env[61947]: INFO nova.compute.manager [-] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Took 1.03 seconds to deallocate network for instance. [ 730.012612] env[61947]: DEBUG nova.compute.claims [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 730.012816] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.084072] env[61947]: DEBUG oslo_concurrency.lockutils [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] Releasing lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.084358] env[61947]: DEBUG nova.compute.manager [req-c085393e-512a-4424-83ea-d2e1f9361453 req-2dc2d5fe-02c6-473e-bc14-5507ae1985ab service nova] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Received event network-vif-deleted-bf2d79c6-1c15-48ac-a876-92e0169f8fa2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.149831] env[61947]: DEBUG nova.scheduler.client.report [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.243029] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.254701] env[61947]: ERROR nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 730.254701] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.254701] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.254701] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.254701] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.254701] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.254701] env[61947]: ERROR nova.compute.manager raise self.value [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.254701] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.254701] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.254701] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.255247] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.255247] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.255247] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 730.255247] env[61947]: ERROR nova.compute.manager [ 730.255247] env[61947]: Traceback (most recent call last): [ 730.255247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.255247] env[61947]: listener.cb(fileno) [ 730.255247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.255247] env[61947]: result = function(*args, **kwargs) [ 730.255247] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.255247] env[61947]: return func(*args, **kwargs) [ 730.255247] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.255247] env[61947]: raise e [ 730.255247] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.255247] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 730.255247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.255247] env[61947]: created_port_ids = self._update_ports_for_instance( [ 730.255247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.255247] env[61947]: with excutils.save_and_reraise_exception(): [ 730.255247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.255247] env[61947]: self.force_reraise() [ 730.255247] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.255247] env[61947]: raise self.value [ 730.255247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.255247] env[61947]: updated_port = self._update_port( [ 730.255247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.255247] env[61947]: _ensure_no_port_binding_failure(port) [ 730.255247] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.255247] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.256097] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 730.256097] env[61947]: Removing descriptor: 17 [ 730.275789] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.276156] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.276393] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.276632] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.276820] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.277007] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.277300] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.277509] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.277715] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.277974] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.278212] env[61947]: DEBUG nova.virt.hardware [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.279134] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891de29e-5dd9-4536-8566-a53fc15e2a06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.287487] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ee3fc7-c75c-480d-a40f-8bba5e1ab0dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.303489] env[61947]: ERROR nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Traceback (most recent call last): [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] yield resources [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.driver.spawn(context, instance, image_meta, [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] vm_ref = self.build_virtual_machine(instance, [ 730.303489] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] for vif in network_info: [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return self._sync_wrapper(fn, *args, **kwargs) [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.wait() [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self[:] = self._gt.wait() [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return self._exit_event.wait() [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.303796] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] current.throw(*self._exc) [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] result = function(*args, **kwargs) [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return func(*args, **kwargs) [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise e [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] nwinfo = self.network_api.allocate_for_instance( [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] created_port_ids = self._update_ports_for_instance( [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] with excutils.save_and_reraise_exception(): [ 730.304121] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.force_reraise() [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise self.value [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] updated_port = self._update_port( [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] _ensure_no_port_binding_failure(port) [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise exception.PortBindingFailed(port_id=port['id']) [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 730.304480] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] [ 730.304480] env[61947]: INFO nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Terminating instance [ 730.306087] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.306282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.306485] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.654984] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.655580] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.658753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.829s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.786869] env[61947]: DEBUG nova.compute.manager [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Received event network-changed-6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.786869] env[61947]: DEBUG nova.compute.manager [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Refreshing instance network info cache due to event network-changed-6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 730.786869] env[61947]: DEBUG oslo_concurrency.lockutils [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] Acquiring lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.824869] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.928011] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.163950] env[61947]: DEBUG nova.compute.utils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.168512] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.168646] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.207736] env[61947]: DEBUG nova.policy [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffaf6398a3a430e8b0996ae478bb324', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd614914b29b544aab77586dc1f274ec9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 731.430650] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.431358] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.431358] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 731.431632] env[61947]: DEBUG oslo_concurrency.lockutils [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] Acquired lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.432352] env[61947]: DEBUG nova.network.neutron [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Refreshing network info cache for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.432850] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f9b1364-5000-423d-9cc7-47cb6fba0192 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.445390] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66dbfea-373f-412a-826c-2f93b3b85bd1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.470105] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c1101f6-4263-47db-bbd6-2be773ea3380 could not be found. [ 731.470333] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 731.470512] env[61947]: INFO nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Took 0.04 seconds to destroy the instance on the hypervisor. [ 731.470743] env[61947]: DEBUG oslo.service.loopingcall [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.477147] env[61947]: DEBUG nova.compute.manager [-] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.477267] env[61947]: DEBUG nova.network.neutron [-] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.503780] env[61947]: DEBUG nova.network.neutron [-] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.542600] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d214ab11-a5bc-43f7-a4c5-5189066edf61 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.550265] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dea45e0-5dfb-46a4-afe3-c0d28eccfcfe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.554268] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Successfully created port: 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.584765] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7de606-db4e-4542-bd26-30ef683e82a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.592314] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90058d4-f31b-4d55-b2b8-a2b34d060223 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.605475] env[61947]: DEBUG nova.compute.provider_tree [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.669413] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.962030] env[61947]: DEBUG nova.network.neutron [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.006275] env[61947]: DEBUG nova.network.neutron [-] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.086515] env[61947]: DEBUG nova.network.neutron [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.110306] env[61947]: DEBUG nova.scheduler.client.report [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.509236] env[61947]: INFO nova.compute.manager [-] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Took 1.03 seconds to deallocate network for instance. [ 732.512130] env[61947]: DEBUG nova.compute.claims [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 732.512317] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.571059] env[61947]: ERROR nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 732.571059] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.571059] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.571059] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.571059] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.571059] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.571059] env[61947]: ERROR nova.compute.manager raise self.value [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.571059] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.571059] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.571059] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.571567] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.571567] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.571567] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 732.571567] env[61947]: ERROR nova.compute.manager [ 732.571567] env[61947]: Traceback (most recent call last): [ 732.571567] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.571567] env[61947]: listener.cb(fileno) [ 732.571567] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.571567] env[61947]: result = function(*args, **kwargs) [ 732.571567] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.571567] env[61947]: return func(*args, **kwargs) [ 732.571567] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.571567] env[61947]: raise e [ 732.571567] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.571567] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 732.571567] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.571567] env[61947]: created_port_ids = self._update_ports_for_instance( [ 732.571567] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.571567] env[61947]: with excutils.save_and_reraise_exception(): [ 732.571567] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.571567] env[61947]: self.force_reraise() [ 732.571567] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.571567] env[61947]: raise self.value [ 732.571567] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.571567] env[61947]: updated_port = self._update_port( [ 732.571567] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.571567] env[61947]: _ensure_no_port_binding_failure(port) [ 732.571567] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.571567] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.572387] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 732.572387] env[61947]: Removing descriptor: 17 [ 732.590991] env[61947]: DEBUG oslo_concurrency.lockutils [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] Releasing lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.591342] env[61947]: DEBUG nova.compute.manager [req-d28d2ac3-a035-43dd-b680-1825a99ca68a req-1bd59775-1932-49ce-b806-2c41d8928930 service nova] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Received event network-vif-deleted-6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.618541] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.619187] env[61947]: ERROR nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Traceback (most recent call last): [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.driver.spawn(context, instance, image_meta, [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] vm_ref = self.build_virtual_machine(instance, [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.619187] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] for vif in network_info: [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return self._sync_wrapper(fn, *args, **kwargs) [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.wait() [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self[:] = self._gt.wait() [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return self._exit_event.wait() [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] current.throw(*self._exc) [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.619569] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] result = function(*args, **kwargs) [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] return func(*args, **kwargs) [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise e [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] nwinfo = self.network_api.allocate_for_instance( [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] created_port_ids = self._update_ports_for_instance( [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] with excutils.save_and_reraise_exception(): [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] self.force_reraise() [ 732.619940] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise self.value [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] updated_port = self._update_port( [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] _ensure_no_port_binding_failure(port) [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] raise exception.PortBindingFailed(port_id=port['id']) [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] nova.exception.PortBindingFailed: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. [ 732.620311] env[61947]: ERROR nova.compute.manager [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] [ 732.620311] env[61947]: DEBUG nova.compute.utils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.621113] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.171s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.622673] env[61947]: INFO nova.compute.claims [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.625484] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Build of instance d58f3c69-ae7a-48d5-a796-34898c6e704c was re-scheduled: Binding failed for port 34ca4adb-5227-4735-99d9-7d7293987aec, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.625907] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.626139] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquiring lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.626284] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Acquired lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.626441] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.679060] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.702575] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.702822] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.702975] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.703169] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.703314] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.703458] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.703655] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.703811] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.703975] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.704146] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.704315] env[61947]: DEBUG nova.virt.hardware [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.705178] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cb7bc8-bf56-4fb0-9e61-40e1d4b5ed22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.713095] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01355ef9-0a41-4761-9eeb-27ea2f7b3dce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.727385] env[61947]: ERROR nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Traceback (most recent call last): [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] yield resources [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.driver.spawn(context, instance, image_meta, [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] vm_ref = self.build_virtual_machine(instance, [ 732.727385] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] for vif in network_info: [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return self._sync_wrapper(fn, *args, **kwargs) [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.wait() [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self[:] = self._gt.wait() [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return self._exit_event.wait() [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.727754] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] current.throw(*self._exc) [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] result = function(*args, **kwargs) [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return func(*args, **kwargs) [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise e [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] nwinfo = self.network_api.allocate_for_instance( [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] created_port_ids = self._update_ports_for_instance( [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] with excutils.save_and_reraise_exception(): [ 732.728145] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.force_reraise() [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise self.value [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] updated_port = self._update_port( [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] _ensure_no_port_binding_failure(port) [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise exception.PortBindingFailed(port_id=port['id']) [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 732.728621] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] [ 732.728621] env[61947]: INFO nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Terminating instance [ 732.729711] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquiring lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.729865] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquired lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.730045] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.813599] env[61947]: DEBUG nova.compute.manager [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Received event network-changed-70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.813858] env[61947]: DEBUG nova.compute.manager [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Refreshing instance network info cache due to event network-changed-70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 732.813992] env[61947]: DEBUG oslo_concurrency.lockutils [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] Acquiring lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.153055] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.221512] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.245414] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.309674] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.724914] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Releasing lock "refresh_cache-d58f3c69-ae7a-48d5-a796-34898c6e704c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.725179] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.725338] env[61947]: DEBUG nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.725506] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.741062] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.812415] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Releasing lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.812816] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.813015] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.815154] env[61947]: DEBUG oslo_concurrency.lockutils [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] Acquired lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.815341] env[61947]: DEBUG nova.network.neutron [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Refreshing network info cache for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.816370] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d3af833-733c-48bb-9c0f-a9686d3a2248 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.826193] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31dca066-39e7-467b-bede-88973a6777ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.851919] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 94d9244e-bf54-442d-8947-71d15d3da551 could not be found. [ 733.852326] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 733.852573] env[61947]: INFO nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.852828] env[61947]: DEBUG oslo.service.loopingcall [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.855817] env[61947]: DEBUG nova.compute.manager [-] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.855924] env[61947]: DEBUG nova.network.neutron [-] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.878038] env[61947]: DEBUG nova.network.neutron [-] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.953220] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1bb5ae-d551-4365-9948-677d2b83477a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.959951] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921ec618-4f88-4a21-b73e-aea942b7dc43 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.991129] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0bd7f1-4a8e-4ba5-b837-b35015fd10b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.998694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c59c7a-92c0-42f2-91ef-4eda9d79cb8d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.011551] env[61947]: DEBUG nova.compute.provider_tree [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.244004] env[61947]: DEBUG nova.network.neutron [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.335245] env[61947]: DEBUG nova.network.neutron [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.380474] env[61947]: DEBUG nova.network.neutron [-] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.404850] env[61947]: DEBUG nova.network.neutron [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.514464] env[61947]: DEBUG nova.scheduler.client.report [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.746318] env[61947]: INFO nova.compute.manager [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] [instance: d58f3c69-ae7a-48d5-a796-34898c6e704c] Took 1.02 seconds to deallocate network for instance. [ 734.884474] env[61947]: INFO nova.compute.manager [-] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Took 1.03 seconds to deallocate network for instance. [ 734.886873] env[61947]: DEBUG nova.compute.claims [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 734.887065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.907663] env[61947]: DEBUG oslo_concurrency.lockutils [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] Releasing lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.907963] env[61947]: DEBUG nova.compute.manager [req-c85f1b19-76b6-43a8-bc7d-5072587fced0 req-5001c1f9-097e-488e-a168-63cbf89a9e1a service nova] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Received event network-vif-deleted-70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 735.019567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.019956] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.022788] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.543s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.024238] env[61947]: INFO nova.compute.claims [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.529279] env[61947]: DEBUG nova.compute.utils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.532174] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.532338] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.571541] env[61947]: DEBUG nova.policy [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '590718317599481d992b2abc69a45c4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9037d85f45f4a25b77218181a1e2359', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 735.782362] env[61947]: INFO nova.scheduler.client.report [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Deleted allocations for instance d58f3c69-ae7a-48d5-a796-34898c6e704c [ 735.902655] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Successfully created port: 12375298-913e-434b-9b9a-d75869d0863f {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.032900] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.290844] env[61947]: DEBUG oslo_concurrency.lockutils [None req-eb32c5d2-cd76-4772-b194-fe9445707be0 tempest-TenantUsagesTestJSON-1167503617 tempest-TenantUsagesTestJSON-1167503617-project-member] Lock "d58f3c69-ae7a-48d5-a796-34898c6e704c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.100s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.426550] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc5d004-bc55-4f67-b3a4-3e878a109b29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.434263] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35231353-3c73-454a-b7a9-bf861a305a25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.465656] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257119e9-6c8c-45d5-bd48-8010a8bc953e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.473352] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41e75a2-ee63-4b97-93f2-e37ba94387d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.486716] env[61947]: DEBUG nova.compute.provider_tree [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.541617] env[61947]: INFO nova.virt.block_device [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Booting with volume b80c8c0b-daad-4fa4-98a6-3a34d8bd6949 at /dev/sda [ 736.602816] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-566cba99-3ce7-435a-8c72-27399577df6f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.611745] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe495160-e94a-436c-b9f5-45a8a4425ab3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.634576] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19507762-767f-4929-9eb8-0d37d6930beb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.642384] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faef83b3-4fc5-48c1-9765-f1e59b06ba30 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.664902] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6a15b7-b002-447e-ae5a-0a0accee0870 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.671529] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b2661f-7389-4c25-ae6b-a019ac16919b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.685077] env[61947]: DEBUG nova.virt.block_device [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating existing volume attachment record: c860c73c-5193-45e2-a0ba-770db4f261ed {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 736.768175] env[61947]: DEBUG nova.compute.manager [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Received event network-changed-12375298-913e-434b-9b9a-d75869d0863f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.768175] env[61947]: DEBUG nova.compute.manager [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Refreshing instance network info cache due to event network-changed-12375298-913e-434b-9b9a-d75869d0863f. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 736.768385] env[61947]: DEBUG oslo_concurrency.lockutils [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] Acquiring lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.768586] env[61947]: DEBUG oslo_concurrency.lockutils [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] Acquired lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.768873] env[61947]: DEBUG nova.network.neutron [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Refreshing network info cache for port 12375298-913e-434b-9b9a-d75869d0863f {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.794705] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.966423] env[61947]: ERROR nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 736.966423] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.966423] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.966423] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.966423] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.966423] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.966423] env[61947]: ERROR nova.compute.manager raise self.value [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.966423] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 736.966423] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.966423] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 736.967013] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.967013] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 736.967013] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 736.967013] env[61947]: ERROR nova.compute.manager [ 736.967013] env[61947]: Traceback (most recent call last): [ 736.967013] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 736.967013] env[61947]: listener.cb(fileno) [ 736.967013] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.967013] env[61947]: result = function(*args, **kwargs) [ 736.967013] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 736.967013] env[61947]: return func(*args, **kwargs) [ 736.967013] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.967013] env[61947]: raise e [ 736.967013] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.967013] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 736.967013] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.967013] env[61947]: created_port_ids = self._update_ports_for_instance( [ 736.967013] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.967013] env[61947]: with excutils.save_and_reraise_exception(): [ 736.967013] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.967013] env[61947]: self.force_reraise() [ 736.967013] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.967013] env[61947]: raise self.value [ 736.967013] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.967013] env[61947]: updated_port = self._update_port( [ 736.967013] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.967013] env[61947]: _ensure_no_port_binding_failure(port) [ 736.967013] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.967013] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 736.968191] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 736.968191] env[61947]: Removing descriptor: 17 [ 736.989423] env[61947]: DEBUG nova.scheduler.client.report [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.290725] env[61947]: DEBUG nova.network.neutron [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.316857] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.439826] env[61947]: DEBUG nova.network.neutron [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.494246] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.494970] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.499300] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.130s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.500934] env[61947]: INFO nova.compute.claims [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.941780] env[61947]: DEBUG oslo_concurrency.lockutils [req-bffc356e-0839-4038-b2af-9156d25cf178 req-2dcbd87e-fdaf-4688-8b22-8bf519e819f4 service nova] Releasing lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.005450] env[61947]: DEBUG nova.compute.utils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.008647] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.008833] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.091234] env[61947]: DEBUG nova.policy [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd85ffd46682f45ef9302c6fd233ab4ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef226aa2daf94b9eaeb3b6b227292ff4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 738.487903] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Successfully created port: 110c363a-3c0d-4d70-99f4-be72dc1e60f0 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.512022] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.808055] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.808055] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.808055] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.808055] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.808348] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.808348] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.808348] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.808348] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.808348] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.809149] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.809846] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.809846] env[61947]: DEBUG nova.virt.hardware [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.810729] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83d66dd-78fb-4971-8479-308c4a21ae8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.824429] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e101fd-6605-4e23-a9ca-e64fc8fe67c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.855143] env[61947]: ERROR nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Traceback (most recent call last): [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] yield resources [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.driver.spawn(context, instance, image_meta, [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] vm_ref = self.build_virtual_machine(instance, [ 738.855143] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] for vif in network_info: [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return self._sync_wrapper(fn, *args, **kwargs) [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.wait() [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self[:] = self._gt.wait() [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return self._exit_event.wait() [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.855524] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] current.throw(*self._exc) [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] result = function(*args, **kwargs) [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return func(*args, **kwargs) [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise e [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] nwinfo = self.network_api.allocate_for_instance( [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] created_port_ids = self._update_ports_for_instance( [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] with excutils.save_and_reraise_exception(): [ 738.855990] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.force_reraise() [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise self.value [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] updated_port = self._update_port( [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] _ensure_no_port_binding_failure(port) [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise exception.PortBindingFailed(port_id=port['id']) [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 738.856392] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] [ 738.856392] env[61947]: INFO nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Terminating instance [ 738.859277] env[61947]: DEBUG nova.compute.manager [req-4c4f56cf-5e1d-411b-a509-4d8031238cb2 req-3fa5225c-81a4-47c5-9b89-50ce528eca16 service nova] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Received event network-vif-deleted-12375298-913e-434b-9b9a-d75869d0863f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 738.859836] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquiring lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.859983] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquired lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.860180] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.936860] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463fb33e-f398-4a22-8c29-7ba52dedcf9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.945298] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad756582-9fdd-41e9-968b-ace1ea855cc3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.982910] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be3ea7d-6445-49d4-ad5c-4973edc031cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.990471] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c38907-de68-43b2-80dc-aaff6f17dd18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.004683] env[61947]: DEBUG nova.compute.provider_tree [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.384052] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.512025] env[61947]: DEBUG nova.scheduler.client.report [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.523700] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.546749] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.561516] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.561744] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.562027] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.564931] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.565812] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.566112] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.566406] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.566643] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.567065] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.567305] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.568328] env[61947]: DEBUG nova.virt.hardware [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.572166] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb518ed3-008e-4d66-9832-ef6e232fad54 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.584026] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c469378-42dc-41d2-9cd6-3dee99ba8060 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.735135] env[61947]: ERROR nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 739.735135] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.735135] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.735135] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.735135] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.735135] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.735135] env[61947]: ERROR nova.compute.manager raise self.value [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.735135] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.735135] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.735135] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.735642] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.735642] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.735642] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 739.735642] env[61947]: ERROR nova.compute.manager [ 739.735642] env[61947]: Traceback (most recent call last): [ 739.735642] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.735642] env[61947]: listener.cb(fileno) [ 739.735642] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.735642] env[61947]: result = function(*args, **kwargs) [ 739.735642] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.735642] env[61947]: return func(*args, **kwargs) [ 739.735642] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.735642] env[61947]: raise e [ 739.735642] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.735642] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 739.735642] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.735642] env[61947]: created_port_ids = self._update_ports_for_instance( [ 739.735642] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.735642] env[61947]: with excutils.save_and_reraise_exception(): [ 739.735642] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.735642] env[61947]: self.force_reraise() [ 739.735642] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.735642] env[61947]: raise self.value [ 739.735642] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.735642] env[61947]: updated_port = self._update_port( [ 739.735642] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.735642] env[61947]: _ensure_no_port_binding_failure(port) [ 739.735642] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.735642] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.736396] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 739.736396] env[61947]: Removing descriptor: 17 [ 739.736396] env[61947]: ERROR nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Traceback (most recent call last): [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] yield resources [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.driver.spawn(context, instance, image_meta, [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.736396] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] vm_ref = self.build_virtual_machine(instance, [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] for vif in network_info: [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self._sync_wrapper(fn, *args, **kwargs) [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.wait() [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self[:] = self._gt.wait() [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self._exit_event.wait() [ 739.736694] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] result = hub.switch() [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self.greenlet.switch() [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] result = function(*args, **kwargs) [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return func(*args, **kwargs) [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise e [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] nwinfo = self.network_api.allocate_for_instance( [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.737013] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] created_port_ids = self._update_ports_for_instance( [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] with excutils.save_and_reraise_exception(): [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.force_reraise() [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise self.value [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] updated_port = self._update_port( [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] _ensure_no_port_binding_failure(port) [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.737339] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise exception.PortBindingFailed(port_id=port['id']) [ 739.737635] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 739.737635] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] [ 739.737635] env[61947]: INFO nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Terminating instance [ 739.740801] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.741370] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.742530] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.015334] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.015961] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.019423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.105s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.019519] env[61947]: DEBUG nova.objects.instance [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61947) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 740.051594] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Releasing lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.051594] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.054758] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa988c84-efb8-4370-a05c-0faaf311068f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.064605] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc7be31-66d8-4e44-aa59-121535d8f861 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.090777] env[61947]: WARNING nova.virt.vmwareapi.driver [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 30b75b32-6fde-4a81-af55-11b51f3e087d could not be found. [ 740.091027] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.091697] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2040fba-6f2d-4068-b771-1b6d1b860f5c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.099330] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b64fff-00ac-4b49-9588-0a3e907cf7f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.125644] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 30b75b32-6fde-4a81-af55-11b51f3e087d could not be found. [ 740.125644] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.125644] env[61947]: INFO nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Took 0.07 seconds to destroy the instance on the hypervisor. [ 740.125644] env[61947]: DEBUG oslo.service.loopingcall [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.125809] env[61947]: DEBUG nova.compute.manager [-] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.125809] env[61947]: DEBUG nova.network.neutron [-] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.155963] env[61947]: DEBUG nova.network.neutron [-] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.261375] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.333018] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.527603] env[61947]: DEBUG nova.compute.utils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.529384] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.529975] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.606400] env[61947]: DEBUG nova.policy [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32a69b3ed4a340dab62d573a0348a0c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d3b8ed5943740ba818f65559043e694', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 740.658458] env[61947]: DEBUG nova.network.neutron [-] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.839933] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.840086] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.840336] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.840665] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4a98adc-f7f3-4bf7-8cb6-ea1099bec959 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.850750] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab9dff9-db73-44c2-9594-bd11c882db42 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.874230] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9 could not be found. [ 740.874506] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.874716] env[61947]: INFO nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.874990] env[61947]: DEBUG oslo.service.loopingcall [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.875249] env[61947]: DEBUG nova.compute.manager [-] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.875362] env[61947]: DEBUG nova.network.neutron [-] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.905084] env[61947]: DEBUG nova.compute.manager [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Received event network-changed-110c363a-3c0d-4d70-99f4-be72dc1e60f0 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.905334] env[61947]: DEBUG nova.compute.manager [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Refreshing instance network info cache due to event network-changed-110c363a-3c0d-4d70-99f4-be72dc1e60f0. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 740.905586] env[61947]: DEBUG oslo_concurrency.lockutils [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] Acquiring lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.905747] env[61947]: DEBUG oslo_concurrency.lockutils [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] Acquired lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.905933] env[61947]: DEBUG nova.network.neutron [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Refreshing network info cache for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.919980] env[61947]: DEBUG nova.network.neutron [-] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.031167] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3c2e5c3-99b0-445e-a376-41151bd555e3 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.031937] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.982s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.035660] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.164521] env[61947]: INFO nova.compute.manager [-] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Took 1.04 seconds to deallocate network for instance. [ 741.177840] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Successfully created port: cc6ffd0c-060d-421f-abab-4be4309bd6b4 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.424694] env[61947]: DEBUG nova.network.neutron [-] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.440406] env[61947]: DEBUG nova.network.neutron [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.562089] env[61947]: DEBUG nova.network.neutron [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.744044] env[61947]: INFO nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Took 0.58 seconds to detach 1 volumes for instance. [ 741.749230] env[61947]: DEBUG nova.compute.claims [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 741.749426] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.932300] env[61947]: INFO nova.compute.manager [-] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Took 1.06 seconds to deallocate network for instance. [ 741.934644] env[61947]: DEBUG nova.compute.claims [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 741.934926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.970872] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c6120f-c3e6-4771-884f-0038f35c8a04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.979931] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1c59d6-c9ef-44b7-91d1-2de41ba39ee8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.022563] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b08bbd-a16a-40a7-aca2-f04df809bb99 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.031696] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352579f5-cbd3-4c12-87de-2f6ac059e4c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.050531] env[61947]: DEBUG nova.compute.provider_tree [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.062038] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.066420] env[61947]: DEBUG oslo_concurrency.lockutils [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] Releasing lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.066773] env[61947]: DEBUG nova.compute.manager [req-4ec8edc6-8d07-48b0-8617-c54197da7d16 req-f0c77055-60c1-465a-9b95-71918a5158e7 service nova] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Received event network-vif-deleted-110c363a-3c0d-4d70-99f4-be72dc1e60f0 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.090877] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.091119] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.091269] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.091447] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.091603] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.091802] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.092018] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.092078] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.092237] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.093244] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.093244] env[61947]: DEBUG nova.virt.hardware [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.093802] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2a8fb6-3cf6-4559-9155-a9534c7c1bdd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.102678] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3878dbd2-88f1-44b1-ab34-fbc76c1bbad7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.554990] env[61947]: DEBUG nova.scheduler.client.report [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.943096] env[61947]: DEBUG nova.compute.manager [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Received event network-changed-cc6ffd0c-060d-421f-abab-4be4309bd6b4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.943263] env[61947]: DEBUG nova.compute.manager [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Refreshing instance network info cache due to event network-changed-cc6ffd0c-060d-421f-abab-4be4309bd6b4. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 742.943503] env[61947]: DEBUG oslo_concurrency.lockutils [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] Acquiring lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.943598] env[61947]: DEBUG oslo_concurrency.lockutils [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] Acquired lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.943753] env[61947]: DEBUG nova.network.neutron [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Refreshing network info cache for port cc6ffd0c-060d-421f-abab-4be4309bd6b4 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 743.032496] env[61947]: ERROR nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 743.032496] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.032496] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.032496] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.032496] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.032496] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.032496] env[61947]: ERROR nova.compute.manager raise self.value [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.032496] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.032496] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.032496] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.033119] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.033119] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.033119] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 743.033119] env[61947]: ERROR nova.compute.manager [ 743.033119] env[61947]: Traceback (most recent call last): [ 743.033119] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.033119] env[61947]: listener.cb(fileno) [ 743.033119] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.033119] env[61947]: result = function(*args, **kwargs) [ 743.033119] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.033119] env[61947]: return func(*args, **kwargs) [ 743.033119] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.033119] env[61947]: raise e [ 743.033119] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.033119] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 743.033119] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.033119] env[61947]: created_port_ids = self._update_ports_for_instance( [ 743.033119] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.033119] env[61947]: with excutils.save_and_reraise_exception(): [ 743.033119] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.033119] env[61947]: self.force_reraise() [ 743.033119] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.033119] env[61947]: raise self.value [ 743.033119] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.033119] env[61947]: updated_port = self._update_port( [ 743.033119] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.033119] env[61947]: _ensure_no_port_binding_failure(port) [ 743.033119] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.033119] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.033844] env[61947]: nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 743.033844] env[61947]: Removing descriptor: 17 [ 743.033844] env[61947]: ERROR nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Traceback (most recent call last): [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] yield resources [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.driver.spawn(context, instance, image_meta, [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.033844] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] vm_ref = self.build_virtual_machine(instance, [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] for vif in network_info: [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self._sync_wrapper(fn, *args, **kwargs) [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.wait() [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self[:] = self._gt.wait() [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self._exit_event.wait() [ 743.034181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] result = hub.switch() [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self.greenlet.switch() [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] result = function(*args, **kwargs) [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return func(*args, **kwargs) [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise e [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] nwinfo = self.network_api.allocate_for_instance( [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.034510] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] created_port_ids = self._update_ports_for_instance( [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] with excutils.save_and_reraise_exception(): [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.force_reraise() [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise self.value [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] updated_port = self._update_port( [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] _ensure_no_port_binding_failure(port) [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.034902] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise exception.PortBindingFailed(port_id=port['id']) [ 743.035238] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 743.035238] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] [ 743.035238] env[61947]: INFO nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Terminating instance [ 743.036637] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.063840] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.032s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.064366] env[61947]: ERROR nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] Traceback (most recent call last): [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.driver.spawn(context, instance, image_meta, [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] vm_ref = self.build_virtual_machine(instance, [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.064366] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] for vif in network_info: [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self._sync_wrapper(fn, *args, **kwargs) [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.wait() [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self[:] = self._gt.wait() [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self._exit_event.wait() [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] result = hub.switch() [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.064689] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return self.greenlet.switch() [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] result = function(*args, **kwargs) [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] return func(*args, **kwargs) [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise e [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] nwinfo = self.network_api.allocate_for_instance( [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] created_port_ids = self._update_ports_for_instance( [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] with excutils.save_and_reraise_exception(): [ 743.065065] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] self.force_reraise() [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise self.value [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] updated_port = self._update_port( [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] _ensure_no_port_binding_failure(port) [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] raise exception.PortBindingFailed(port_id=port['id']) [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] nova.exception.PortBindingFailed: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. [ 743.065425] env[61947]: ERROR nova.compute.manager [instance: 947b697d-4120-491d-83ce-21681974c224] [ 743.065884] env[61947]: DEBUG nova.compute.utils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.068770] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.100s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.069319] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Build of instance 947b697d-4120-491d-83ce-21681974c224 was re-scheduled: Binding failed for port a07c729e-d13d-46ab-b727-03b40b2b586c, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.069907] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.070330] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquiring lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.070488] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Acquired lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.070657] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.464249] env[61947]: DEBUG nova.network.neutron [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.666137] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.772076] env[61947]: DEBUG nova.network.neutron [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.805150] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.962968] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f064aa-8a4c-440b-a377-867029c2ebf7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.971179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddee8b6a-d911-4700-adfa-57c2de5c2e2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.009019] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9133cc03-c201-4d59-ba81-4cf752cf0f20 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.017334] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7ebc74-5519-46ad-afe0-f90ea230ed14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.031451] env[61947]: DEBUG nova.compute.provider_tree [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.274842] env[61947]: DEBUG oslo_concurrency.lockutils [req-0752d501-29de-4318-ae8d-4a12a1330f2f req-f9659f7b-d7fc-4fc7-bd01-1ea28344f16b service nova] Releasing lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.275289] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "74e03575-297e-4e08-9236-98d8be80b546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.275377] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.275576] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquired lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.275758] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.308222] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Releasing lock "refresh_cache-947b697d-4120-491d-83ce-21681974c224" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.308450] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.308624] env[61947]: DEBUG nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.308786] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.323938] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.534859] env[61947]: DEBUG nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.799245] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.826144] env[61947]: DEBUG nova.network.neutron [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.882456] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.975881] env[61947]: DEBUG nova.compute.manager [req-23586a30-8e33-4b84-ab37-170f17c0149b req-ce13b673-5c85-4f50-9900-38000f7c7333 service nova] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Received event network-vif-deleted-cc6ffd0c-060d-421f-abab-4be4309bd6b4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.040461] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.041119] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Traceback (most recent call last): [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.driver.spawn(context, instance, image_meta, [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] vm_ref = self.build_virtual_machine(instance, [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.041119] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] for vif in network_info: [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return self._sync_wrapper(fn, *args, **kwargs) [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.wait() [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self[:] = self._gt.wait() [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return self._exit_event.wait() [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] current.throw(*self._exc) [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.041445] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] result = function(*args, **kwargs) [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] return func(*args, **kwargs) [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise e [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] nwinfo = self.network_api.allocate_for_instance( [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] created_port_ids = self._update_ports_for_instance( [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] with excutils.save_and_reraise_exception(): [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] self.force_reraise() [ 745.042051] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise self.value [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] updated_port = self._update_port( [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] _ensure_no_port_binding_failure(port) [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] raise exception.PortBindingFailed(port_id=port['id']) [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] nova.exception.PortBindingFailed: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. [ 745.042463] env[61947]: ERROR nova.compute.manager [instance: c8c66631-1fae-4a03-810f-66d64a75b118] [ 745.042463] env[61947]: DEBUG nova.compute.utils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.043538] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.491s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.043752] env[61947]: DEBUG nova.objects.instance [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lazy-loading 'resources' on Instance uuid 32fe176d-ebc2-46a4-8e93-3b2555a681d4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 745.046030] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Build of instance c8c66631-1fae-4a03-810f-66d64a75b118 was re-scheduled: Binding failed for port c4e7f3c1-4609-498a-8f56-d8bdbf5bc1cb, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.046364] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.046559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.046708] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.046895] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.330058] env[61947]: INFO nova.compute.manager [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] [instance: 947b697d-4120-491d-83ce-21681974c224] Took 1.02 seconds to deallocate network for instance. [ 745.386049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Releasing lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.386049] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.386049] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 745.386236] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30f3315f-366f-46f2-b13f-bd1b3e28cf73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.395622] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd76527a-4aa8-4c23-b7de-063235cda604 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.418099] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d4b9119-f375-45a0-99c6-ceb75f43c776 could not be found. [ 745.418342] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 745.418538] env[61947]: INFO nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.418774] env[61947]: DEBUG oslo.service.loopingcall [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.419014] env[61947]: DEBUG nova.compute.manager [-] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.419099] env[61947]: DEBUG nova.network.neutron [-] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.440703] env[61947]: DEBUG nova.network.neutron [-] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.566537] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.646342] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.852084] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20514f80-9d7e-42a8-8f79-2794c8e40a4c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.860102] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f554fd02-4ae5-444e-94c9-d63f256220a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.892115] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57753cb-e7e4-48ac-aaa0-a166dc18daf0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.900041] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c7ceba-4704-4904-b6eb-329f61d631f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.914643] env[61947]: DEBUG nova.compute.provider_tree [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.942734] env[61947]: DEBUG nova.network.neutron [-] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.149393] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-c8c66631-1fae-4a03-810f-66d64a75b118" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.149631] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.149809] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.149975] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.163811] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.360113] env[61947]: INFO nova.scheduler.client.report [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Deleted allocations for instance 947b697d-4120-491d-83ce-21681974c224 [ 746.420888] env[61947]: DEBUG nova.scheduler.client.report [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.447352] env[61947]: INFO nova.compute.manager [-] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Took 1.03 seconds to deallocate network for instance. [ 746.450161] env[61947]: DEBUG nova.compute.claims [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 746.450334] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.666651] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.868574] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2c14346f-170a-4514-8ad1-49ab16fe0ed2 tempest-ServerActionsTestOtherA-1326277279 tempest-ServerActionsTestOtherA-1326277279-project-member] Lock "947b697d-4120-491d-83ce-21681974c224" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.339s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.929076] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.929596] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.437s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.947264] env[61947]: INFO nova.scheduler.client.report [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleted allocations for instance 32fe176d-ebc2-46a4-8e93-3b2555a681d4 [ 747.170471] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: c8c66631-1fae-4a03-810f-66d64a75b118] Took 1.02 seconds to deallocate network for instance. [ 747.371027] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.456419] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e429060c-78fb-4757-99c4-860fe3640a6a tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "32fe176d-ebc2-46a4-8e93-3b2555a681d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.614s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.726083] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c38c4b-0dd5-41d4-b322-5ee6028afab1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.733763] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c46ebef-eee4-49ed-9a22-475bc0fd7e45 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.764616] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f624dc7a-e5e2-4e91-897b-b071d72e0086 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.771696] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176c28b2-b45c-4f4e-ba65-20e2f31f2f6b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.784443] env[61947]: DEBUG nova.compute.provider_tree [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.890995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.914009] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.914307] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.914586] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.914772] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.914939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.918042] env[61947]: INFO nova.compute.manager [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Terminating instance [ 747.921669] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.921937] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquired lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.921998] env[61947]: DEBUG nova.network.neutron [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.206887] env[61947]: INFO nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted allocations for instance c8c66631-1fae-4a03-810f-66d64a75b118 [ 748.288025] env[61947]: DEBUG nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.449381] env[61947]: DEBUG nova.network.neutron [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.499628] env[61947]: DEBUG nova.network.neutron [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.718282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "c8c66631-1fae-4a03-810f-66d64a75b118" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.138s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.796516] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.797137] env[61947]: ERROR nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Traceback (most recent call last): [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.driver.spawn(context, instance, image_meta, [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] vm_ref = self.build_virtual_machine(instance, [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.797137] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] for vif in network_info: [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return self._sync_wrapper(fn, *args, **kwargs) [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.wait() [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self[:] = self._gt.wait() [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return self._exit_event.wait() [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] current.throw(*self._exc) [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.797517] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] result = function(*args, **kwargs) [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] return func(*args, **kwargs) [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise e [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] nwinfo = self.network_api.allocate_for_instance( [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] created_port_ids = self._update_ports_for_instance( [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] with excutils.save_and_reraise_exception(): [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] self.force_reraise() [ 748.797983] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise self.value [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] updated_port = self._update_port( [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] _ensure_no_port_binding_failure(port) [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] raise exception.PortBindingFailed(port_id=port['id']) [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] nova.exception.PortBindingFailed: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. [ 748.798375] env[61947]: ERROR nova.compute.manager [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] [ 748.798375] env[61947]: DEBUG nova.compute.utils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.799032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.786s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.801835] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Build of instance 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37 was re-scheduled: Binding failed for port da42e119-9f9c-4444-8733-fef0693a8338, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 748.802249] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 748.802466] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.802607] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.802763] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.001986] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Releasing lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.002545] env[61947]: DEBUG nova.compute.manager [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 749.002783] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 749.003869] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc4a63b-f15a-440f-9f15-d5091548f1d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.014367] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 749.014626] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1c0bb17-5189-4687-a9ea-41ae5a4988c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.020993] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 749.020993] env[61947]: value = "task-1224190" [ 749.020993] env[61947]: _type = "Task" [ 749.020993] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.032185] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.220431] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.327755] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.421818] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.530733] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224190, 'name': PowerOffVM_Task, 'duration_secs': 0.102594} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.533305] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 749.533305] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 749.533880] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42143806-dea3-4bef-87fb-e31240397168 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.560265] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 749.561413] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 749.561413] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleting the datastore file [datastore2] 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 749.563405] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1ce8f58-e95c-4fdd-a38b-a526d92b2fad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.570604] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for the task: (returnval){ [ 749.570604] env[61947]: value = "task-1224192" [ 749.570604] env[61947]: _type = "Task" [ 749.570604] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.578752] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.664807] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ac884c-3881-4b38-9b52-1f8ccd4e0115 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.672266] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1433ce-3320-4a44-b547-0eed2933fa61 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.702886] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4477385c-a6cd-4780-82a3-f71f404111ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.711062] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3634adc0-affc-4870-8d85-faafa8aaf966 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.726478] env[61947]: DEBUG nova.compute.provider_tree [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.739791] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.927242] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.927649] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 749.927968] env[61947]: DEBUG nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.928353] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.944026] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.082054] env[61947]: DEBUG oslo_vmware.api [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Task: {'id': task-1224192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086055} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.082054] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 750.082054] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 750.082054] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 750.082054] env[61947]: INFO nova.compute.manager [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Took 1.08 seconds to destroy the instance on the hypervisor. [ 750.082400] env[61947]: DEBUG oslo.service.loopingcall [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.082567] env[61947]: DEBUG nova.compute.manager [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.082662] env[61947]: DEBUG nova.network.neutron [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.097090] env[61947]: DEBUG nova.network.neutron [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.234062] env[61947]: DEBUG nova.scheduler.client.report [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.446589] env[61947]: DEBUG nova.network.neutron [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.600050] env[61947]: DEBUG nova.network.neutron [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.740105] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.740783] env[61947]: ERROR nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Traceback (most recent call last): [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.driver.spawn(context, instance, image_meta, [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] vm_ref = self.build_virtual_machine(instance, [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.740783] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] for vif in network_info: [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return self._sync_wrapper(fn, *args, **kwargs) [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.wait() [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self[:] = self._gt.wait() [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return self._exit_event.wait() [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] current.throw(*self._exc) [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.741145] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] result = function(*args, **kwargs) [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] return func(*args, **kwargs) [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise e [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] nwinfo = self.network_api.allocate_for_instance( [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] created_port_ids = self._update_ports_for_instance( [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] with excutils.save_and_reraise_exception(): [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] self.force_reraise() [ 750.741519] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise self.value [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] updated_port = self._update_port( [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] _ensure_no_port_binding_failure(port) [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] raise exception.PortBindingFailed(port_id=port['id']) [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] nova.exception.PortBindingFailed: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. [ 750.741886] env[61947]: ERROR nova.compute.manager [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] [ 750.741886] env[61947]: DEBUG nova.compute.utils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 750.743044] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Build of instance 8a4df3f1-67f3-404d-953f-caa6ba1b1641 was re-scheduled: Binding failed for port bf2d79c6-1c15-48ac-a876-92e0169f8fa2, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 750.743465] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 750.743690] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.743835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.743985] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.744913] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.233s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.949748] env[61947]: INFO nova.compute.manager [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37] Took 1.02 seconds to deallocate network for instance. [ 751.102801] env[61947]: INFO nova.compute.manager [-] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Took 1.02 seconds to deallocate network for instance. [ 751.271395] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.364350] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.545017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d380143-9363-4fd5-85be-9ec25e81c1ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.551314] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7a7002-3367-481a-9e82-7f4437d618cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.581630] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56af3226-e3d6-4ca6-aadc-b0aa85195051 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.588625] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23af25fe-34df-4d7f-87a4-e4ffa21a12e1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.601546] env[61947]: DEBUG nova.compute.provider_tree [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.609202] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.870027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-8a4df3f1-67f3-404d-953f-caa6ba1b1641" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.870027] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 751.870027] env[61947]: DEBUG nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.870027] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.885445] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.980160] env[61947]: INFO nova.scheduler.client.report [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted allocations for instance 86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37 [ 752.105161] env[61947]: DEBUG nova.scheduler.client.report [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.388456] env[61947]: DEBUG nova.network.neutron [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.488528] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7735fa13-0187-43d8-97fb-8f8e7cde6d78 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "86d9ab25-b3d6-4cc4-8b96-2ed2b3959f37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.850s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.610707] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.865s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.610707] env[61947]: ERROR nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Traceback (most recent call last): [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.driver.spawn(context, instance, image_meta, [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.610707] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] vm_ref = self.build_virtual_machine(instance, [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] for vif in network_info: [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return self._sync_wrapper(fn, *args, **kwargs) [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.wait() [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self[:] = self._gt.wait() [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return self._exit_event.wait() [ 752.611213] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] current.throw(*self._exc) [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] result = function(*args, **kwargs) [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] return func(*args, **kwargs) [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise e [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] nwinfo = self.network_api.allocate_for_instance( [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] created_port_ids = self._update_ports_for_instance( [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.611600] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] with excutils.save_and_reraise_exception(): [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] self.force_reraise() [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise self.value [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] updated_port = self._update_port( [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] _ensure_no_port_binding_failure(port) [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] raise exception.PortBindingFailed(port_id=port['id']) [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] nova.exception.PortBindingFailed: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. [ 752.611977] env[61947]: ERROR nova.compute.manager [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] [ 752.612364] env[61947]: DEBUG nova.compute.utils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.612539] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.725s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.617037] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Build of instance 8c1101f6-4263-47db-bbd6-2be773ea3380 was re-scheduled: Binding failed for port 6ab9f2f2-448e-4f5b-ac7f-d12fd598cc9f, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.617037] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.617037] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.617037] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.617263] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.891540] env[61947]: INFO nova.compute.manager [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 8a4df3f1-67f3-404d-953f-caa6ba1b1641] Took 1.02 seconds to deallocate network for instance. [ 752.991565] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.140289] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.262467] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.410936] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a2d179-1ab8-487c-8fa7-49bd49358e41 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.421160] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197ffcbe-17f1-4750-8f4e-7d9bb25eec56 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.451133] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7407e92-5e10-4727-ab62-b831e844b442 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.458940] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218b7156-3e3e-4ae1-a06c-8631fc8eae79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.473316] env[61947]: DEBUG nova.compute.provider_tree [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.510947] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.765716] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-8c1101f6-4263-47db-bbd6-2be773ea3380" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.766283] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.766283] env[61947]: DEBUG nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.766415] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.790467] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.922078] env[61947]: INFO nova.scheduler.client.report [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted allocations for instance 8a4df3f1-67f3-404d-953f-caa6ba1b1641 [ 753.975615] env[61947]: DEBUG nova.scheduler.client.report [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.088630] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.088909] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.117802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.118103] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.297115] env[61947]: DEBUG nova.network.neutron [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.433534] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c0facde2-4f57-496b-8e5f-011cdc3bcec7 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "8a4df3f1-67f3-404d-953f-caa6ba1b1641" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.052s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.479847] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.480505] env[61947]: ERROR nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Traceback (most recent call last): [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.driver.spawn(context, instance, image_meta, [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] vm_ref = self.build_virtual_machine(instance, [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.480505] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] for vif in network_info: [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return self._sync_wrapper(fn, *args, **kwargs) [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.wait() [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self[:] = self._gt.wait() [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return self._exit_event.wait() [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] current.throw(*self._exc) [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.480837] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] result = function(*args, **kwargs) [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] return func(*args, **kwargs) [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise e [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] nwinfo = self.network_api.allocate_for_instance( [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] created_port_ids = self._update_ports_for_instance( [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] with excutils.save_and_reraise_exception(): [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] self.force_reraise() [ 754.481172] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise self.value [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] updated_port = self._update_port( [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] _ensure_no_port_binding_failure(port) [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] raise exception.PortBindingFailed(port_id=port['id']) [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] nova.exception.PortBindingFailed: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. [ 754.481803] env[61947]: ERROR nova.compute.manager [instance: 94d9244e-bf54-442d-8947-71d15d3da551] [ 754.481803] env[61947]: DEBUG nova.compute.utils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.482355] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.166s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.483798] env[61947]: INFO nova.compute.claims [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.487611] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Build of instance 94d9244e-bf54-442d-8947-71d15d3da551 was re-scheduled: Binding failed for port 70b8b082-4ba4-4b5e-ad0a-4ce105d61e5d, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 754.487611] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 754.487876] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquiring lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.487876] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Acquired lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.488113] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.798960] env[61947]: INFO nova.compute.manager [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 8c1101f6-4263-47db-bbd6-2be773ea3380] Took 1.03 seconds to deallocate network for instance. [ 754.938727] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.016069] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.112073] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.460736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.615795] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Releasing lock "refresh_cache-94d9244e-bf54-442d-8947-71d15d3da551" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.616026] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 755.616221] env[61947]: DEBUG nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.616388] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.641841] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.834054] env[61947]: INFO nova.scheduler.client.report [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocations for instance 8c1101f6-4263-47db-bbd6-2be773ea3380 [ 755.858909] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd3a942-658c-4269-b872-d3e537ca78c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.869472] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4328ec51-22e1-494c-b9ec-ff0e3e8f4c7c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.902658] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6948f4cb-62b4-4cc2-bdd9-ea64d059d59e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.911375] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cc2359-cb33-4236-86dc-0bf43b8e8725 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.924396] env[61947]: DEBUG nova.compute.provider_tree [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.144956] env[61947]: DEBUG nova.network.neutron [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.342190] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b114306e-8a2e-4ed6-9e4e-bae2c1cc688a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "8c1101f6-4263-47db-bbd6-2be773ea3380" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.564s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.427046] env[61947]: DEBUG nova.scheduler.client.report [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.648424] env[61947]: INFO nova.compute.manager [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] [instance: 94d9244e-bf54-442d-8947-71d15d3da551] Took 1.03 seconds to deallocate network for instance. [ 756.845284] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.933841] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.933841] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.937487] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.187s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.094302] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.094574] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.232531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.233134] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.365069] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.441672] env[61947]: DEBUG nova.compute.utils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.445989] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.446196] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.492243] env[61947]: DEBUG nova.policy [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32a69b3ed4a340dab62d573a0348a0c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d3b8ed5943740ba818f65559043e694', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 757.683522] env[61947]: INFO nova.scheduler.client.report [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Deleted allocations for instance 94d9244e-bf54-442d-8947-71d15d3da551 [ 757.773916] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d442fd-dd06-4091-8701-f9b89519af16 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.784623] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6629aee-ff78-4892-8455-05ce2b016572 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.817451] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Successfully created port: 4f13b041-13bb-46c1-a43d-eaec3cd9a63e {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.818280] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed422be7-658a-4284-97b7-ba80e78278d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.826986] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf953b8-fb52-4eb7-bee2-3cb10837011b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.838422] env[61947]: DEBUG nova.compute.provider_tree [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.948635] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.197444] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a9cb6713-7e2c-4a2b-8b44-d60addb30b42 tempest-ServerActionsTestJSON-342339161 tempest-ServerActionsTestJSON-342339161-project-member] Lock "94d9244e-bf54-442d-8947-71d15d3da551" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.991s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.346030] env[61947]: DEBUG nova.scheduler.client.report [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.575102] env[61947]: DEBUG nova.compute.manager [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Received event network-changed-4f13b041-13bb-46c1-a43d-eaec3cd9a63e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.575102] env[61947]: DEBUG nova.compute.manager [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Refreshing instance network info cache due to event network-changed-4f13b041-13bb-46c1-a43d-eaec3cd9a63e. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 758.575102] env[61947]: DEBUG oslo_concurrency.lockutils [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] Acquiring lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.575102] env[61947]: DEBUG oslo_concurrency.lockutils [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] Acquired lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.575102] env[61947]: DEBUG nova.network.neutron [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Refreshing network info cache for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.701499] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.848837] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.849748] env[61947]: ERROR nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Traceback (most recent call last): [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.driver.spawn(context, instance, image_meta, [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] vm_ref = self.build_virtual_machine(instance, [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.849748] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] for vif in network_info: [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return self._sync_wrapper(fn, *args, **kwargs) [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.wait() [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self[:] = self._gt.wait() [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return self._exit_event.wait() [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] current.throw(*self._exc) [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.850372] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] result = function(*args, **kwargs) [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] return func(*args, **kwargs) [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise e [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] nwinfo = self.network_api.allocate_for_instance( [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] created_port_ids = self._update_ports_for_instance( [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] with excutils.save_and_reraise_exception(): [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] self.force_reraise() [ 758.850888] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise self.value [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] updated_port = self._update_port( [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] _ensure_no_port_binding_failure(port) [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] raise exception.PortBindingFailed(port_id=port['id']) [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] nova.exception.PortBindingFailed: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. [ 758.851290] env[61947]: ERROR nova.compute.manager [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] [ 758.851290] env[61947]: DEBUG nova.compute.utils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.854973] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.920s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.858016] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Build of instance 30b75b32-6fde-4a81-af55-11b51f3e087d was re-scheduled: Binding failed for port 12375298-913e-434b-9b9a-d75869d0863f, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 758.859627] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 758.859863] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquiring lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.860024] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Acquired lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.860192] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.951838] env[61947]: ERROR nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 758.951838] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.951838] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.951838] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.951838] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.951838] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.951838] env[61947]: ERROR nova.compute.manager raise self.value [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.951838] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.951838] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.951838] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.952490] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.952490] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.952490] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 758.952490] env[61947]: ERROR nova.compute.manager [ 758.952490] env[61947]: Traceback (most recent call last): [ 758.952490] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.952490] env[61947]: listener.cb(fileno) [ 758.952490] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.952490] env[61947]: result = function(*args, **kwargs) [ 758.952490] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.952490] env[61947]: return func(*args, **kwargs) [ 758.952490] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.952490] env[61947]: raise e [ 758.952490] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.952490] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 758.952490] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.952490] env[61947]: created_port_ids = self._update_ports_for_instance( [ 758.952490] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.952490] env[61947]: with excutils.save_and_reraise_exception(): [ 758.952490] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.952490] env[61947]: self.force_reraise() [ 758.952490] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.952490] env[61947]: raise self.value [ 758.952490] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.952490] env[61947]: updated_port = self._update_port( [ 758.952490] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.952490] env[61947]: _ensure_no_port_binding_failure(port) [ 758.952490] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.952490] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.953228] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 758.953228] env[61947]: Removing descriptor: 18 [ 758.957874] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.992660] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.993360] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.993574] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.993762] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.994062] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.994242] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.995287] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.995481] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.995660] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.995849] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.996077] env[61947]: DEBUG nova.virt.hardware [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.996964] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac02405c-c7bd-4fda-a59f-8b6a96bd94ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.007614] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4caf6d-630e-47cc-900c-ade92c46fa7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.026293] env[61947]: ERROR nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Traceback (most recent call last): [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] yield resources [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.driver.spawn(context, instance, image_meta, [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] vm_ref = self.build_virtual_machine(instance, [ 759.026293] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] for vif in network_info: [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return self._sync_wrapper(fn, *args, **kwargs) [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.wait() [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self[:] = self._gt.wait() [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return self._exit_event.wait() [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.026785] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] current.throw(*self._exc) [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] result = function(*args, **kwargs) [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return func(*args, **kwargs) [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise e [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] nwinfo = self.network_api.allocate_for_instance( [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] created_port_ids = self._update_ports_for_instance( [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] with excutils.save_and_reraise_exception(): [ 759.027225] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.force_reraise() [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise self.value [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] updated_port = self._update_port( [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] _ensure_no_port_binding_failure(port) [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise exception.PortBindingFailed(port_id=port['id']) [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 759.027598] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] [ 759.027598] env[61947]: INFO nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Terminating instance [ 759.028890] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.103233] env[61947]: DEBUG nova.network.neutron [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.190146] env[61947]: DEBUG nova.network.neutron [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.226131] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.379657] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.446220] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.673391] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a14926-3af6-4d96-a73d-80eb61b196ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.683091] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b2a31c-58e3-4e89-b59e-ec39a947364c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.714395] env[61947]: DEBUG oslo_concurrency.lockutils [req-94a87676-c842-47dd-82a2-adbaf17f3566 req-953ddfbb-4ca1-4694-b55e-bb453eeaa4c1 service nova] Releasing lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.715019] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquired lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.719364] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.721067] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd70af2-341f-4c6c-9d9a-b61d5c3d8fcc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.734449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbaf584-0537-4eab-bc56-741cd9a0bf55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.751313] env[61947]: DEBUG nova.compute.provider_tree [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.948554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Releasing lock "refresh_cache-30b75b32-6fde-4a81-af55-11b51f3e087d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.948805] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 759.949034] env[61947]: DEBUG nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.949359] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.969205] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.246839] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.254694] env[61947]: DEBUG nova.scheduler.client.report [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.357995] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.472066] env[61947]: DEBUG nova.network.neutron [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.598987] env[61947]: DEBUG nova.compute.manager [req-f822dbce-5e61-4827-bc09-64ba8c25781f req-3fbb75b8-de7e-46d9-9d73-8e825929b5d5 service nova] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Received event network-vif-deleted-4f13b041-13bb-46c1-a43d-eaec3cd9a63e {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 760.762668] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.763346] env[61947]: ERROR nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Traceback (most recent call last): [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.driver.spawn(context, instance, image_meta, [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] vm_ref = self.build_virtual_machine(instance, [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.763346] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] for vif in network_info: [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self._sync_wrapper(fn, *args, **kwargs) [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.wait() [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self[:] = self._gt.wait() [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self._exit_event.wait() [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] result = hub.switch() [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.763772] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return self.greenlet.switch() [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] result = function(*args, **kwargs) [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] return func(*args, **kwargs) [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise e [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] nwinfo = self.network_api.allocate_for_instance( [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] created_port_ids = self._update_ports_for_instance( [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] with excutils.save_and_reraise_exception(): [ 760.764158] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] self.force_reraise() [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise self.value [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] updated_port = self._update_port( [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] _ensure_no_port_binding_failure(port) [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] raise exception.PortBindingFailed(port_id=port['id']) [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] nova.exception.PortBindingFailed: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. [ 760.764530] env[61947]: ERROR nova.compute.manager [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] [ 760.764846] env[61947]: DEBUG nova.compute.utils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.765204] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.315s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.767993] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Build of instance 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9 was re-scheduled: Binding failed for port 110c363a-3c0d-4d70-99f4-be72dc1e60f0, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.769333] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.769333] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.769333] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.769333] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.862027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Releasing lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.862027] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.862211] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.862545] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1af440e8-8577-4b2b-b43b-1edd869e2c02 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.872658] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38c8e87-0f07-4055-9502-e217640328be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.895239] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 46dbb96b-dc8b-4fc9-bc54-181c5a308916 could not be found. [ 760.895474] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.895666] env[61947]: INFO nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Took 0.03 seconds to destroy the instance on the hypervisor. [ 760.896044] env[61947]: DEBUG oslo.service.loopingcall [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.896290] env[61947]: DEBUG nova.compute.manager [-] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.896384] env[61947]: DEBUG nova.network.neutron [-] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.911869] env[61947]: DEBUG nova.network.neutron [-] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.974532] env[61947]: INFO nova.compute.manager [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] [instance: 30b75b32-6fde-4a81-af55-11b51f3e087d] Took 1.03 seconds to deallocate network for instance. [ 761.299261] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.370740] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.414366] env[61947]: DEBUG nova.network.neutron [-] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.573843] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5f2aef-5008-4fc9-b86f-ebc63b1e3307 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.581121] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcf4b5f-1c31-461a-848f-cfddf9286aa8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.610733] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e573d7-f61b-4672-8a8b-18ea0d0d9a2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.617311] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9d3ca6-2609-4bcd-b920-cc4e2921b274 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.630707] env[61947]: DEBUG nova.compute.provider_tree [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.723845] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.724090] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.874191] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.874469] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.874699] env[61947]: DEBUG nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.874806] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.893921] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.917294] env[61947]: INFO nova.compute.manager [-] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Took 1.02 seconds to deallocate network for instance. [ 761.918867] env[61947]: DEBUG nova.compute.claims [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 761.919086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.003227] env[61947]: INFO nova.scheduler.client.report [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Deleted allocations for instance 30b75b32-6fde-4a81-af55-11b51f3e087d [ 762.133632] env[61947]: DEBUG nova.scheduler.client.report [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.396374] env[61947]: DEBUG nova.network.neutron [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.514329] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c4b49b7-6edc-4bc0-8944-7d1e4b766b35 tempest-ServersTestBootFromVolume-1284205491 tempest-ServersTestBootFromVolume-1284205491-project-member] Lock "30b75b32-6fde-4a81-af55-11b51f3e087d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.702s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.638906] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.639786] env[61947]: ERROR nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Traceback (most recent call last): [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.driver.spawn(context, instance, image_meta, [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] vm_ref = self.build_virtual_machine(instance, [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.639786] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] for vif in network_info: [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self._sync_wrapper(fn, *args, **kwargs) [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.wait() [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self[:] = self._gt.wait() [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self._exit_event.wait() [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] result = hub.switch() [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.640181] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return self.greenlet.switch() [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] result = function(*args, **kwargs) [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] return func(*args, **kwargs) [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise e [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] nwinfo = self.network_api.allocate_for_instance( [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] created_port_ids = self._update_ports_for_instance( [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] with excutils.save_and_reraise_exception(): [ 762.640589] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] self.force_reraise() [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise self.value [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] updated_port = self._update_port( [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] _ensure_no_port_binding_failure(port) [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] raise exception.PortBindingFailed(port_id=port['id']) [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] nova.exception.PortBindingFailed: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. [ 762.640981] env[61947]: ERROR nova.compute.manager [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] [ 762.641321] env[61947]: DEBUG nova.compute.utils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.641782] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.751s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.643251] env[61947]: INFO nova.compute.claims [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.646759] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Build of instance 3d4b9119-f375-45a0-99c6-ceb75f43c776 was re-scheduled: Binding failed for port cc6ffd0c-060d-421f-abab-4be4309bd6b4, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 762.647260] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 762.647488] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.647634] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquired lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.647792] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.902487] env[61947]: INFO nova.compute.manager [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9] Took 1.03 seconds to deallocate network for instance. [ 762.953483] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.953670] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 763.017522] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.166434] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.258247] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.458715] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 763.458960] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 763.459101] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 763.547508] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.759362] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Releasing lock "refresh_cache-3d4b9119-f375-45a0-99c6-ceb75f43c776" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.759499] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 763.759672] env[61947]: DEBUG nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.759840] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.776555] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.945153] env[61947]: INFO nova.scheduler.client.report [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Deleted allocations for instance 4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9 [ 763.963785] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 763.964147] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 764.004106] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee90a2d-26f5-4fec-9115-deee8014ef64 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.017332] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c056af-3e33-43d7-ab53-c3662e5fd2c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.021733] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.022124] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.022346] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 764.022622] env[61947]: DEBUG nova.objects.instance [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lazy-loading 'info_cache' on Instance uuid 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 764.054021] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5ffce1-ea0b-4271-a4f3-64ddd919c503 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.061480] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22a7554-4acc-49ed-bff7-b93b2049de4f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.076999] env[61947]: DEBUG nova.compute.provider_tree [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.281283] env[61947]: DEBUG nova.network.neutron [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.454842] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e904334-bd8d-44cd-8419-4a865072a3cf tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "4558e0ba-8bf4-4ee7-a16b-9e4929a21ab9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.775s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.579873] env[61947]: DEBUG nova.scheduler.client.report [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.783884] env[61947]: INFO nova.compute.manager [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 3d4b9119-f375-45a0-99c6-ceb75f43c776] Took 1.02 seconds to deallocate network for instance. [ 764.956927] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.048824] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.086707] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.087386] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.090936] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.351s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.092462] env[61947]: INFO nova.compute.claims [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.475094] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.597620] env[61947]: DEBUG nova.compute.utils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.601445] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 765.610294] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.811233] env[61947]: INFO nova.scheduler.client.report [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Deleted allocations for instance 3d4b9119-f375-45a0-99c6-ceb75f43c776 [ 765.821228] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "72818e40-624b-4c04-888b-bb622f7f96d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.821600] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.102327] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.112367] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.112367] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 766.112367] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.112550] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.112585] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.112741] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.112894] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.113587] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.113587] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 766.113587] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.323030] env[61947]: DEBUG oslo_concurrency.lockutils [None req-185f71da-f257-40ef-ba6c-a2ce478d65df tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "3d4b9119-f375-45a0-99c6-ceb75f43c776" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.171s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.363605] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637907dc-b0d2-49cf-b7a5-d6fc652ad025 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.371446] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cb453f-a995-4198-92d0-4a56735bbeb1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.401694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc4027f-bbb5-419b-abf8-cf2fa2dee598 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.409269] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e685d13a-8005-46e8-9d0d-b323664628b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.422351] env[61947]: DEBUG nova.compute.provider_tree [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.616511] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.826664] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.925653] env[61947]: DEBUG nova.scheduler.client.report [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.115045] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.140535] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.140788] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.140944] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.141138] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.141372] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.141551] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.141760] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.141916] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.142651] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.142651] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.142988] env[61947]: DEBUG nova.virt.hardware [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.143740] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b426838-008f-49c1-b4ed-535b77f7ea95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.152322] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a3c20c-0229-4916-a7aa-37393d69d4f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.165467] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.171258] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Creating folder: Project (a28457b38f184348bcd1e94dde80f231). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 767.171553] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e26a5a8-375e-458e-a0bb-ee856c84b092 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.184262] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Created folder: Project (a28457b38f184348bcd1e94dde80f231) in parent group-v264556. [ 767.184384] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Creating folder: Instances. Parent ref: group-v264573. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 767.184622] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9913fde0-5a2b-43e7-a6af-5d456c7bb76f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.193246] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Created folder: Instances in parent group-v264573. [ 767.193416] env[61947]: DEBUG oslo.service.loopingcall [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.193613] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 767.193807] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e695331c-470c-4329-ac0f-872da6a0ec33 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.209887] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.209887] env[61947]: value = "task-1224196" [ 767.209887] env[61947]: _type = "Task" [ 767.209887] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.217081] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224196, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.354334] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.431666] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.432230] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.435303] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.826s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.435303] env[61947]: DEBUG nova.objects.instance [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lazy-loading 'resources' on Instance uuid 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.720688] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224196, 'name': CreateVM_Task, 'duration_secs': 0.241665} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.720860] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 767.721286] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.721453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.721757] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 767.722013] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06685e8e-4d7b-476c-8cbd-be224f048f9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.727021] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 767.727021] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ac8ccb-e43d-d96d-e7ee-880cbc8aac80" [ 767.727021] env[61947]: _type = "Task" [ 767.727021] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.734748] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ac8ccb-e43d-d96d-e7ee-880cbc8aac80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.940792] env[61947]: DEBUG nova.compute.utils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.942298] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.942467] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.036654] env[61947]: DEBUG nova.policy [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bade4ae175344af7b925933ee1ef015b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55da0357c2b94db6bfc69b12bba49092', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 768.240265] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ac8ccb-e43d-d96d-e7ee-880cbc8aac80, 'name': SearchDatastore_Task, 'duration_secs': 0.008579} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.240769] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.241233] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.242264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.242264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.242264] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.242389] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5db98598-3ba4-4af6-b43e-3111844e0008 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.253740] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.253953] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 768.254872] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5dcab0-3fca-4733-94fa-815f6ac8d113 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.262089] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 768.262089] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f9120-2ec0-ffc0-344e-52e4b4ea642c" [ 768.262089] env[61947]: _type = "Task" [ 768.262089] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.269832] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f9120-2ec0-ffc0-344e-52e4b4ea642c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.273578] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6269b94d-3bcd-480b-9a07-e0e972df7a1f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.279898] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773e52bb-494b-4d46-ba1e-5af1e8df6ab6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.312124] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a49f9a5-5018-413c-a1a2-9c334ab998d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.319869] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6816dea-d9b7-4cc6-8993-9e698628f3c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.333402] env[61947]: DEBUG nova.compute.provider_tree [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.449890] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.465133] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Successfully created port: 6170ba23-bb1e-42cd-8ccc-23ef1df38a96 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.777270] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f9120-2ec0-ffc0-344e-52e4b4ea642c, 'name': SearchDatastore_Task, 'duration_secs': 0.008741} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.778239] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6615ad1-9f00-4636-8c24-46cf6a2ee300 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.785210] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 768.785210] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522b3489-9e01-b0d1-76e7-896d0bdd0dda" [ 768.785210] env[61947]: _type = "Task" [ 768.785210] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.793753] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522b3489-9e01-b0d1-76e7-896d0bdd0dda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.838099] env[61947]: DEBUG nova.scheduler.client.report [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.215548] env[61947]: DEBUG nova.compute.manager [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Received event network-changed-6170ba23-bb1e-42cd-8ccc-23ef1df38a96 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.215709] env[61947]: DEBUG nova.compute.manager [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Refreshing instance network info cache due to event network-changed-6170ba23-bb1e-42cd-8ccc-23ef1df38a96. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 769.215926] env[61947]: DEBUG oslo_concurrency.lockutils [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] Acquiring lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.216261] env[61947]: DEBUG oslo_concurrency.lockutils [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] Acquired lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.216477] env[61947]: DEBUG nova.network.neutron [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Refreshing network info cache for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.299755] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522b3489-9e01-b0d1-76e7-896d0bdd0dda, 'name': SearchDatastore_Task, 'duration_secs': 0.008731} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.300049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.300297] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 769.300558] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5c89f2a-0df5-44de-bb19-78ba45f75ef4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.308211] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 769.308211] env[61947]: value = "task-1224197" [ 769.308211] env[61947]: _type = "Task" [ 769.308211] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.326278] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.343301] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.346427] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.835s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.351336] env[61947]: INFO nova.compute.claims [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.377656] env[61947]: INFO nova.scheduler.client.report [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Deleted allocations for instance 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8 [ 769.444412] env[61947]: ERROR nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 769.444412] env[61947]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.444412] env[61947]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.444412] env[61947]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.444412] env[61947]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.444412] env[61947]: ERROR nova.compute.manager self.force_reraise() [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.444412] env[61947]: ERROR nova.compute.manager raise self.value [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.444412] env[61947]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.444412] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.444412] env[61947]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.445094] env[61947]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.445094] env[61947]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.445094] env[61947]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 769.445094] env[61947]: ERROR nova.compute.manager [ 769.445094] env[61947]: Traceback (most recent call last): [ 769.445094] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.445094] env[61947]: listener.cb(fileno) [ 769.445094] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.445094] env[61947]: result = function(*args, **kwargs) [ 769.445094] env[61947]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.445094] env[61947]: return func(*args, **kwargs) [ 769.445094] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.445094] env[61947]: raise e [ 769.445094] env[61947]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.445094] env[61947]: nwinfo = self.network_api.allocate_for_instance( [ 769.445094] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.445094] env[61947]: created_port_ids = self._update_ports_for_instance( [ 769.445094] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.445094] env[61947]: with excutils.save_and_reraise_exception(): [ 769.445094] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.445094] env[61947]: self.force_reraise() [ 769.445094] env[61947]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.445094] env[61947]: raise self.value [ 769.445094] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.445094] env[61947]: updated_port = self._update_port( [ 769.445094] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.445094] env[61947]: _ensure_no_port_binding_failure(port) [ 769.445094] env[61947]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.445094] env[61947]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.447443] env[61947]: nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 769.447443] env[61947]: Removing descriptor: 18 [ 769.460616] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.488320] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.488601] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.488753] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.488932] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.489206] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.489445] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.490239] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.490239] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.490239] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.490373] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.490612] env[61947]: DEBUG nova.virt.hardware [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.491559] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34034ded-7de7-4ba8-923a-3c55c3cbac09 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.500352] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316f5363-7c7a-4dfc-bb8e-a9736fe752cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.515034] env[61947]: ERROR nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Traceback (most recent call last): [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] yield resources [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.driver.spawn(context, instance, image_meta, [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] vm_ref = self.build_virtual_machine(instance, [ 769.515034] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] for vif in network_info: [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return self._sync_wrapper(fn, *args, **kwargs) [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.wait() [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self[:] = self._gt.wait() [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return self._exit_event.wait() [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.515752] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] current.throw(*self._exc) [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] result = function(*args, **kwargs) [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return func(*args, **kwargs) [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise e [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] nwinfo = self.network_api.allocate_for_instance( [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] created_port_ids = self._update_ports_for_instance( [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] with excutils.save_and_reraise_exception(): [ 769.516326] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.force_reraise() [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise self.value [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] updated_port = self._update_port( [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] _ensure_no_port_binding_failure(port) [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise exception.PortBindingFailed(port_id=port['id']) [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 769.516919] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] [ 769.516919] env[61947]: INFO nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Terminating instance [ 769.519168] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.747775] env[61947]: DEBUG nova.network.neutron [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.805965] env[61947]: DEBUG nova.network.neutron [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.818302] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438659} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.818734] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 769.819154] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.819582] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69ce49f8-d01b-43d0-929e-4563a7f36e1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.829050] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 769.829050] env[61947]: value = "task-1224198" [ 769.829050] env[61947]: _type = "Task" [ 769.829050] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.835318] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224198, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.887614] env[61947]: DEBUG oslo_concurrency.lockutils [None req-074802aa-6985-443f-87a9-f555e302aea1 tempest-ServerShowV247Test-1160637535 tempest-ServerShowV247Test-1160637535-project-member] Lock "0c7a6d23-7196-4af2-9f52-6e5a09f26fd8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.973s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.312870] env[61947]: DEBUG oslo_concurrency.lockutils [req-25d6cffb-cc60-4a20-9ae5-48d7365384b5 req-4cd8024d-ad2e-40c7-9db8-6597b0bb377f service nova] Releasing lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.313318] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.313508] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.336343] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224198, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074047} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.336597] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.337365] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6a7c0c-7d8b-460a-8dd3-28b20ddb3dec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.357310] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.357866] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cd08bc4-717a-4688-aa23-7511bbbfca11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.380800] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 770.380800] env[61947]: value = "task-1224199" [ 770.380800] env[61947]: _type = "Task" [ 770.380800] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.392936] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224199, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.669119] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69e1d34-25b0-46c4-80a5-7740a00bda49 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.678811] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802cbb51-a240-4d51-9ebd-edb2f25e538b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.708175] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313ad6cf-6463-4f8a-92c1-b3e910b5151c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.715323] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a91bb0-dc64-4502-94df-5a50709a4d67 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.728346] env[61947]: DEBUG nova.compute.provider_tree [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.837177] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.891540] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224199, 'name': ReconfigVM_Task, 'duration_secs': 0.27176} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.891754] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.892390] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e3ce129-66e2-4bbd-9fac-b221a870aed0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.894705] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.901038] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 770.901038] env[61947]: value = "task-1224200" [ 770.901038] env[61947]: _type = "Task" [ 770.901038] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.911670] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224200, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.231294] env[61947]: DEBUG nova.scheduler.client.report [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.239307] env[61947]: DEBUG nova.compute.manager [req-f3b5931a-b96e-487e-ab7e-6baf542fa5ae req-34d4b90c-dead-46b4-a470-1927918f4f1b service nova] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Received event network-vif-deleted-6170ba23-bb1e-42cd-8ccc-23ef1df38a96 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.397343] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Releasing lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.397775] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.398010] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 771.398323] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c569b52-e310-4712-a831-57f221d11759 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.410412] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdd0aeb-824e-4f34-97b8-27ded9e41aae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.423317] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224200, 'name': Rename_Task, 'duration_secs': 0.139586} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.423889] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 771.424160] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a1af2f1-fca2-4e12-af2e-e2848bf1ca15 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.429624] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 771.429624] env[61947]: value = "task-1224201" [ 771.429624] env[61947]: _type = "Task" [ 771.429624] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.436239] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be1fdafa-3c4d-430c-99c5-202e6ccebf26 could not be found. [ 771.436440] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.436617] env[61947]: INFO nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Took 0.04 seconds to destroy the instance on the hypervisor. [ 771.436921] env[61947]: DEBUG oslo.service.loopingcall [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.437417] env[61947]: DEBUG nova.compute.manager [-] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.437531] env[61947]: DEBUG nova.network.neutron [-] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.441625] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.455011] env[61947]: DEBUG nova.network.neutron [-] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.739057] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.739680] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.743545] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.282s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.744998] env[61947]: INFO nova.compute.claims [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.940709] env[61947]: DEBUG oslo_vmware.api [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224201, 'name': PowerOnVM_Task, 'duration_secs': 0.403627} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.941420] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 771.941772] env[61947]: INFO nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Took 4.83 seconds to spawn the instance on the hypervisor. [ 771.942460] env[61947]: DEBUG nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 771.943581] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b311f8-4806-4ac5-84ed-e3258ba26d68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.957847] env[61947]: DEBUG nova.network.neutron [-] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.251541] env[61947]: DEBUG nova.compute.utils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.253398] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.253398] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.304449] env[61947]: DEBUG nova.policy [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bade4ae175344af7b925933ee1ef015b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55da0357c2b94db6bfc69b12bba49092', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 772.462100] env[61947]: INFO nova.compute.manager [-] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Took 1.02 seconds to deallocate network for instance. [ 772.462100] env[61947]: INFO nova.compute.manager [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Took 24.59 seconds to build instance. [ 772.464498] env[61947]: DEBUG nova.compute.claims [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Aborting claim: {{(pid=61947) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 772.464721] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.730055] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Successfully created port: 0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.758024] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.968145] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7f2cafcd-164b-4120-b6f8-852a9f17d030 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.157s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.145865] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5d8d0e-99b6-4fec-8100-40080de6fca4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.154495] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbaad4e-601c-435c-8c7d-67811571f401 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.187012] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10535755-1382-4de8-a32c-a2068fae2d5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.194648] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f1a827-f27f-4619-92a6-a7ec33e03b7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.207907] env[61947]: DEBUG nova.compute.provider_tree [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.475136] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.571422] env[61947]: INFO nova.compute.manager [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Rebuilding instance [ 773.622209] env[61947]: DEBUG nova.compute.manager [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.623096] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10746166-386c-4764-b875-424f0f8d046a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.710524] env[61947]: DEBUG nova.scheduler.client.report [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.770018] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.798259] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.799016] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.799016] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.799689] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.799978] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.800319] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.800683] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.800990] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.807644] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.807644] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.807644] env[61947]: DEBUG nova.virt.hardware [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.807644] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055fed08-af52-4335-b358-b48e429489e0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.813267] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a39988-4b75-4d9f-aeb3-4ed8101489f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.997631] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.136463] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 774.136929] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-870bceab-46a6-4639-8ab5-d0d929ae6349 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.146428] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 774.146428] env[61947]: value = "task-1224202" [ 774.146428] env[61947]: _type = "Task" [ 774.146428] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.158049] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.216553] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.216767] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.220166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.855s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.221882] env[61947]: INFO nova.compute.claims [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.305808] env[61947]: DEBUG nova.compute.manager [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Received event network-vif-plugged-0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 774.306065] env[61947]: DEBUG oslo_concurrency.lockutils [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] Acquiring lock "683d40c9-b738-486a-806a-7e895637ee85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.306274] env[61947]: DEBUG oslo_concurrency.lockutils [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] Lock "683d40c9-b738-486a-806a-7e895637ee85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.306435] env[61947]: DEBUG oslo_concurrency.lockutils [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] Lock "683d40c9-b738-486a-806a-7e895637ee85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.306617] env[61947]: DEBUG nova.compute.manager [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] No waiting events found dispatching network-vif-plugged-0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.306748] env[61947]: WARNING nova.compute.manager [req-6cfba60d-2ef2-4470-bdec-7659ad548ddf req-7cba39a6-a8fa-4e66-8845-f73ba9ed2a1d service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Received unexpected event network-vif-plugged-0d86518d-eb0e-4c5b-996b-c88ec7987305 for instance with vm_state building and task_state spawning. [ 774.426039] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Successfully updated port: 0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.656327] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224202, 'name': PowerOffVM_Task, 'duration_secs': 0.154865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.656608] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 774.656808] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 774.657575] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbd8053-f33e-472b-b5ce-f1033a1d6bab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.663975] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 774.664199] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba60a42a-e59f-4029-895c-5a1983b83d13 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.686010] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 774.686234] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 774.686407] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Deleting the datastore file [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.686664] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aff39343-ce0b-4fee-a3ad-019f7dc7c15c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.693009] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 774.693009] env[61947]: value = "task-1224204" [ 774.693009] env[61947]: _type = "Task" [ 774.693009] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.700496] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.726338] env[61947]: DEBUG nova.compute.utils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.730030] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.730030] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.811021] env[61947]: DEBUG nova.policy [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47cf366a472d4fd5b15f464c7b22d1ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbe056f2420d4781a1923c7af9c28570', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 774.929523] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.929523] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.929523] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.041271] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.041474] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.207314] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097697} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.207711] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.207902] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 775.208113] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 775.216671] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Successfully created port: 224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.230798] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.482447] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.577906] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa82a270-e5d1-439b-927f-44fa01a65ce6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.585438] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd9a312-3a4d-4a1c-9362-3a8ed287ee69 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.620247] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c483f4-e3fc-496b-95e0-881946cc436b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.628417] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9e4fde-c0ee-4622-a9e0-b4cd484249a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.642220] env[61947]: DEBUG nova.compute.provider_tree [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.719013] env[61947]: DEBUG nova.network.neutron [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Updating instance_info_cache with network_info: [{"id": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "address": "fa:16:3e:d2:6f:77", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.17", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d86518d-eb", "ovs_interfaceid": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.145370] env[61947]: DEBUG nova.scheduler.client.report [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.222174] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Releasing lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.222174] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Instance network_info: |[{"id": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "address": "fa:16:3e:d2:6f:77", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.17", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d86518d-eb", "ovs_interfaceid": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.222597] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:6f:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d86518d-eb0e-4c5b-996b-c88ec7987305', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.230662] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Creating folder: Project (55da0357c2b94db6bfc69b12bba49092). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.231277] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-863ea791-537d-4a66-8a2d-29c4a20326ac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.245023] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.245023] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.245023] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.245023] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.245281] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.245281] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.245281] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.245281] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.245281] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.245456] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.245456] env[61947]: DEBUG nova.virt.hardware [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.245757] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.247898] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09e1fb9-b422-4bec-bd58-7310396e0130 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.251560] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Created folder: Project (55da0357c2b94db6bfc69b12bba49092) in parent group-v264556. [ 776.251733] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Creating folder: Instances. Parent ref: group-v264576. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.252374] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d925c24a-6761-4071-9be0-c8e7dd9c6398 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.257324] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde8cccf-28ab-4eac-b039-649e515d4761 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.261929] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Created folder: Instances in parent group-v264576. [ 776.262162] env[61947]: DEBUG oslo.service.loopingcall [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.264123] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 776.265029] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42b4d975-f3aa-4856-95ce-335a41695eca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.280320] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.280522] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.280676] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.280851] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.280994] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.281155] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.281355] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.281547] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.281667] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.281821] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.281985] env[61947]: DEBUG nova.virt.hardware [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.290144] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3561108-662f-4a1f-a6f4-11582016837c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.292571] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.297918] env[61947]: DEBUG oslo.service.loopingcall [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.299027] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 776.299402] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82ef9ea6-5d14-43f3-8f1a-75db28d46053 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.314351] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.314351] env[61947]: value = "task-1224207" [ 776.314351] env[61947]: _type = "Task" [ 776.314351] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.316699] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a897e3e-29e8-49c3-b7d0-fa5ece34d17a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.341743] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224207, 'name': CreateVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.341950] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.341950] env[61947]: value = "task-1224208" [ 776.341950] env[61947]: _type = "Task" [ 776.341950] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.343142] env[61947]: DEBUG nova.compute.manager [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Received event network-changed-0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.343324] env[61947]: DEBUG nova.compute.manager [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Refreshing instance network info cache due to event network-changed-0d86518d-eb0e-4c5b-996b-c88ec7987305. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 776.343529] env[61947]: DEBUG oslo_concurrency.lockutils [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] Acquiring lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.343668] env[61947]: DEBUG oslo_concurrency.lockutils [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] Acquired lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.343823] env[61947]: DEBUG nova.network.neutron [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Refreshing network info cache for port 0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.353252] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224208, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.654027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.654027] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.654960] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.429s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.656717] env[61947]: INFO nova.compute.claims [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.829108] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224207, 'name': CreateVM_Task, 'duration_secs': 0.287962} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.829108] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.832361] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Successfully updated port: 224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.839190] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.839190] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.839190] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.839190] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5784c91-5dd6-46d9-860f-d9578e9dd227 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.842929] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 776.842929] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]529ebc4c-a8e3-8cc8-b215-b2fa748fdcb8" [ 776.842929] env[61947]: _type = "Task" [ 776.842929] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.860489] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]529ebc4c-a8e3-8cc8-b215-b2fa748fdcb8, 'name': SearchDatastore_Task, 'duration_secs': 0.009886} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.861586] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.861826] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.862060] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.862316] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.866017] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.866017] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224208, 'name': CreateVM_Task, 'duration_secs': 0.244164} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.866017] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddac3a56-16a8-4b42-99b6-4d4771e7eed0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.866017] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.866017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.866017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.866488] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.866488] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-205d213c-fc9d-4ee2-8634-7cb3ade39193 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.870397] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 776.870397] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526de098-b153-4761-ce9f-0f2db7ab2f95" [ 776.870397] env[61947]: _type = "Task" [ 776.870397] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.874712] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.874880] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 776.876945] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c2792d9-7565-497c-a401-70bbf305f84f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.881179] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526de098-b153-4761-ce9f-0f2db7ab2f95, 'name': SearchDatastore_Task, 'duration_secs': 0.007385} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.881717] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.881932] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.882140] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.883873] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 776.883873] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52646904-1b9a-9e20-f1d3-2199da0d2a61" [ 776.883873] env[61947]: _type = "Task" [ 776.883873] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.893297] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52646904-1b9a-9e20-f1d3-2199da0d2a61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.061215] env[61947]: DEBUG nova.network.neutron [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Updated VIF entry in instance network info cache for port 0d86518d-eb0e-4c5b-996b-c88ec7987305. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 777.061568] env[61947]: DEBUG nova.network.neutron [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Updating instance_info_cache with network_info: [{"id": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "address": "fa:16:3e:d2:6f:77", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.17", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d86518d-eb", "ovs_interfaceid": "0d86518d-eb0e-4c5b-996b-c88ec7987305", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.163566] env[61947]: DEBUG nova.compute.utils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.166163] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.166163] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.204026] env[61947]: DEBUG nova.policy [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a665ac33d2b4e2486ce324baf006115', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8aa679df5f264a438984f000b16f7d82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 777.336358] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.336522] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquired lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.336683] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.398164] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52646904-1b9a-9e20-f1d3-2199da0d2a61, 'name': SearchDatastore_Task, 'duration_secs': 0.007619} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.398992] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c2874a2-1849-4b90-a786-10b18f6104fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.404689] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 777.404689] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5258cdba-2624-12f4-d95e-579d71545dc3" [ 777.404689] env[61947]: _type = "Task" [ 777.404689] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.412489] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5258cdba-2624-12f4-d95e-579d71545dc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.473097] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Successfully created port: a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.564557] env[61947]: DEBUG oslo_concurrency.lockutils [req-7f967d35-d044-4ceb-8c74-76635c38fd55 req-3bbf9b76-61c0-42c1-8da9-51e838e8e49b service nova] Releasing lock "refresh_cache-683d40c9-b738-486a-806a-7e895637ee85" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.669387] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.875620] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.920212] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5258cdba-2624-12f4-d95e-579d71545dc3, 'name': SearchDatastore_Task, 'duration_secs': 0.008985} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.922430] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.923020] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 683d40c9-b738-486a-806a-7e895637ee85/683d40c9-b738-486a-806a-7e895637ee85.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.923020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.923171] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.923383] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29bb2078-6508-45b3-9dc4-d78de2a0c2b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.928085] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a02607fe-8155-4dcd-a20f-f3f55c7bc5b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.935325] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 777.935325] env[61947]: value = "task-1224209" [ 777.935325] env[61947]: _type = "Task" [ 777.935325] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.939546] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.939770] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 777.942853] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff4c66bd-22e9-46f5-8dc0-c98a4f899cf1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.949331] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.954490] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 777.954490] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a10d45-60eb-1f8f-0b18-7fc4823a1803" [ 777.954490] env[61947]: _type = "Task" [ 777.954490] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.959252] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a10d45-60eb-1f8f-0b18-7fc4823a1803, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.005010] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d419df-ba9b-4a69-9964-49d26b154fbe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.012456] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb2745a-7966-46de-bd17-d5c88b3c3006 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.042633] env[61947]: DEBUG nova.network.neutron [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updating instance_info_cache with network_info: [{"id": "224bbfc2-7511-45fe-95b2-caf904d6794b", "address": "fa:16:3e:0c:ff:29", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.30", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224bbfc2-75", "ovs_interfaceid": "224bbfc2-7511-45fe-95b2-caf904d6794b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.044550] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93ef982-a5f8-48be-acaf-06fd1f3840f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.051923] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f15b177-2ad8-45df-bb0e-b2cdfea7da6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.068150] env[61947]: DEBUG nova.compute.provider_tree [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.382609] env[61947]: DEBUG nova.compute.manager [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Received event network-vif-plugged-224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.382841] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Acquiring lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.383072] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.383247] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.383412] env[61947]: DEBUG nova.compute.manager [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] No waiting events found dispatching network-vif-plugged-224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.383571] env[61947]: WARNING nova.compute.manager [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Received unexpected event network-vif-plugged-224bbfc2-7511-45fe-95b2-caf904d6794b for instance with vm_state building and task_state spawning. [ 778.383728] env[61947]: DEBUG nova.compute.manager [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Received event network-changed-224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.383884] env[61947]: DEBUG nova.compute.manager [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Refreshing instance network info cache due to event network-changed-224bbfc2-7511-45fe-95b2-caf904d6794b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.384058] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Acquiring lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.445464] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43632} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.445664] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 683d40c9-b738-486a-806a-7e895637ee85/683d40c9-b738-486a-806a-7e895637ee85.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 778.445876] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.446163] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a26dbe0b-6125-4852-b523-4d659d3b0470 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.452137] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 778.452137] env[61947]: value = "task-1224210" [ 778.452137] env[61947]: _type = "Task" [ 778.452137] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.462827] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224210, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.466272] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a10d45-60eb-1f8f-0b18-7fc4823a1803, 'name': SearchDatastore_Task, 'duration_secs': 0.008123} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.466979] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3530d9e1-19ee-400f-bd8a-a6c20cdd0741 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.471902] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 778.471902] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266938f-3e64-fa9a-cec7-b727146bd35d" [ 778.471902] env[61947]: _type = "Task" [ 778.471902] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.479077] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266938f-3e64-fa9a-cec7-b727146bd35d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.548681] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Releasing lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.549039] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Instance network_info: |[{"id": "224bbfc2-7511-45fe-95b2-caf904d6794b", "address": "fa:16:3e:0c:ff:29", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.30", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224bbfc2-75", "ovs_interfaceid": "224bbfc2-7511-45fe-95b2-caf904d6794b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.549394] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Acquired lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.549575] env[61947]: DEBUG nova.network.neutron [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Refreshing network info cache for port 224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.550946] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:ff:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '224bbfc2-7511-45fe-95b2-caf904d6794b', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.558455] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Creating folder: Project (dbe056f2420d4781a1923c7af9c28570). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.559406] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c29838bb-b527-4a29-bec5-5e0ba610cd45 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.569151] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Created folder: Project (dbe056f2420d4781a1923c7af9c28570) in parent group-v264556. [ 778.569325] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Creating folder: Instances. Parent ref: group-v264580. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.569538] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1aabb7bc-5885-4b99-9dd9-cdb4359814da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.571522] env[61947]: DEBUG nova.scheduler.client.report [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.580669] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Created folder: Instances in parent group-v264580. [ 778.580882] env[61947]: DEBUG oslo.service.loopingcall [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.581066] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.581255] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-112f45a6-b40c-4aab-82d0-c28e20137376 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.599533] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.599533] env[61947]: value = "task-1224213" [ 778.599533] env[61947]: _type = "Task" [ 778.599533] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.607310] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224213, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.685484] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.710431] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.710716] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.710901] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.711128] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.711347] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.711543] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.711789] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.711982] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.712211] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.712413] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.712618] env[61947]: DEBUG nova.virt.hardware [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.713630] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62455b75-ef37-4cd2-9f7e-ff246d784ed8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.722117] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb9e3d3-aa72-4f0f-afd2-f2b28a72650c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.968035] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224210, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057161} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.968035] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.968035] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e92293-1599-42ca-9d8e-cdc1a2d2071f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.987442] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 683d40c9-b738-486a-806a-7e895637ee85/683d40c9-b738-486a-806a-7e895637ee85.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.992315] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c5a9da-3b5d-43f7-b905-73e5e85103b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.011447] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266938f-3e64-fa9a-cec7-b727146bd35d, 'name': SearchDatastore_Task, 'duration_secs': 0.008779} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.012655] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.012915] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 779.013238] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 779.013238] env[61947]: value = "task-1224214" [ 779.013238] env[61947]: _type = "Task" [ 779.013238] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.013426] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-111b8caf-e15f-4890-86c9-c2bcf1584d5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.024168] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224214, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.025411] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 779.025411] env[61947]: value = "task-1224215" [ 779.025411] env[61947]: _type = "Task" [ 779.025411] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.032911] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.075928] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.077031] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.079244] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.160s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.110811] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224213, 'name': CreateVM_Task, 'duration_secs': 0.291438} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.111071] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.111727] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.111887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.112234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.112538] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-360c10f9-8eff-4b78-b71a-8f39ac381e38 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.117294] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 779.117294] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267be6e-478b-01df-b4c6-b2179c3a1991" [ 779.117294] env[61947]: _type = "Task" [ 779.117294] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.127661] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267be6e-478b-01df-b4c6-b2179c3a1991, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.301234] env[61947]: DEBUG nova.compute.manager [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Received event network-vif-plugged-a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.301234] env[61947]: DEBUG oslo_concurrency.lockutils [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] Acquiring lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.301324] env[61947]: DEBUG oslo_concurrency.lockutils [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.301525] env[61947]: DEBUG oslo_concurrency.lockutils [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.301635] env[61947]: DEBUG nova.compute.manager [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] No waiting events found dispatching network-vif-plugged-a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 779.301795] env[61947]: WARNING nova.compute.manager [req-519baaec-5e4e-44cf-b548-accd88f7d3f1 req-65e17836-b711-4c5b-89c3-180f2f9d47b1 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Received unexpected event network-vif-plugged-a67cfa09-3b87-4186-8afb-19040cd3dcea for instance with vm_state building and task_state spawning. [ 779.414537] env[61947]: DEBUG nova.network.neutron [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updated VIF entry in instance network info cache for port 224bbfc2-7511-45fe-95b2-caf904d6794b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 779.415692] env[61947]: DEBUG nova.network.neutron [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updating instance_info_cache with network_info: [{"id": "224bbfc2-7511-45fe-95b2-caf904d6794b", "address": "fa:16:3e:0c:ff:29", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.30", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224bbfc2-75", "ovs_interfaceid": "224bbfc2-7511-45fe-95b2-caf904d6794b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.525279] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224214, 'name': ReconfigVM_Task, 'duration_secs': 0.266589} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.525564] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 683d40c9-b738-486a-806a-7e895637ee85/683d40c9-b738-486a-806a-7e895637ee85.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.526328] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96b50a77-62ce-47cc-b0df-ef80731ad9bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.537998] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224215, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.537998] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 779.537998] env[61947]: value = "task-1224216" [ 779.537998] env[61947]: _type = "Task" [ 779.537998] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.544568] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224216, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.584134] env[61947]: DEBUG nova.compute.utils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.589436] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.589436] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.628454] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267be6e-478b-01df-b4c6-b2179c3a1991, 'name': SearchDatastore_Task, 'duration_secs': 0.033098} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.628759] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.629032] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.629308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.629453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.629631] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.629905] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dd58581-e57e-43d5-9c6e-cca029c0a10e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.634133] env[61947]: DEBUG nova.policy [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41d4a2d39ebc4ce0a0e76350f907e7df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58b83ea896bd4654aaa9e81f635a6a55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 779.643875] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.644122] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.644929] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cca2e3ff-cf7e-46e2-a141-895cab4587d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.654422] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 779.654422] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b6fd-f515-8213-4f04-1cf4c04e2e28" [ 779.654422] env[61947]: _type = "Task" [ 779.654422] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.662581] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b6fd-f515-8213-4f04-1cf4c04e2e28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.769649] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Successfully updated port: a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.917837] env[61947]: DEBUG oslo_concurrency.lockutils [req-16670f54-b462-4a16-ad88-bdbad73db220 req-dae5f7d1-225e-476f-ba14-3b72f98fc6f1 service nova] Releasing lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.940142] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefcf887-6692-4d35-aa4c-f550ef88fee8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.946853] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266450dc-5c81-4bc6-a117-16475d696ca4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.981683] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24457978-d4da-442c-98bc-70cff0a5fde5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.988979] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d587972a-9303-436b-9695-ba31660daf68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.001982] env[61947]: DEBUG nova.compute.provider_tree [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.031172] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Successfully created port: c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.037638] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224215, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.878962} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.037955] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 780.038201] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.041394] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a65baee-9e17-4440-b956-16621cd7b961 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.048857] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224216, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.049050] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 780.049050] env[61947]: value = "task-1224217" [ 780.049050] env[61947]: _type = "Task" [ 780.049050] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.055902] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224217, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.090227] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.168443] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b6fd-f515-8213-4f04-1cf4c04e2e28, 'name': SearchDatastore_Task, 'duration_secs': 0.055783} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.169277] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dd73dd3-1fe7-4dd8-99ad-6b77cb68cc82 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.174438] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 780.174438] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522d906d-2154-3770-fa06-24da01ab2982" [ 780.174438] env[61947]: _type = "Task" [ 780.174438] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.182110] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522d906d-2154-3770-fa06-24da01ab2982, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.274307] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.274307] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquired lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.274307] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.413401] env[61947]: DEBUG nova.compute.manager [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Received event network-changed-a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.413584] env[61947]: DEBUG nova.compute.manager [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Refreshing instance network info cache due to event network-changed-a67cfa09-3b87-4186-8afb-19040cd3dcea. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 780.413800] env[61947]: DEBUG oslo_concurrency.lockutils [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] Acquiring lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.504744] env[61947]: DEBUG nova.scheduler.client.report [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.548090] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224216, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.556969] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224217, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060411} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.557201] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.557894] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74ebd93-0af2-4424-945d-15f01c4d5459 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.576634] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.576870] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc640ff2-238e-4a35-9c68-fb121dbed0cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.595178] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 780.595178] env[61947]: value = "task-1224218" [ 780.595178] env[61947]: _type = "Task" [ 780.595178] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.605978] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224218, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.684297] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522d906d-2154-3770-fa06-24da01ab2982, 'name': SearchDatastore_Task, 'duration_secs': 0.008776} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.684566] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.684855] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] d642bdb3-8e98-4667-9af7-71dae739e6b1/d642bdb3-8e98-4667-9af7-71dae739e6b1.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.685137] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81a7d0cc-e133-4b6d-b357-cdad351190be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.691718] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 780.691718] env[61947]: value = "task-1224219" [ 780.691718] env[61947]: _type = "Task" [ 780.691718] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.699716] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.808361] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.979737] env[61947]: DEBUG nova.network.neutron [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Updating instance_info_cache with network_info: [{"id": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "address": "fa:16:3e:12:ba:7d", "network": {"id": "fdb2f6d6-01b5-4ea3-8b54-6ddb4f4f366d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1736584182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8aa679df5f264a438984f000b16f7d82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa67cfa09-3b", "ovs_interfaceid": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.012784] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.013666] env[61947]: ERROR nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Traceback (most recent call last): [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.driver.spawn(context, instance, image_meta, [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] vm_ref = self.build_virtual_machine(instance, [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.013666] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] for vif in network_info: [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return self._sync_wrapper(fn, *args, **kwargs) [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.wait() [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self[:] = self._gt.wait() [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return self._exit_event.wait() [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] current.throw(*self._exc) [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.014108] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] result = function(*args, **kwargs) [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] return func(*args, **kwargs) [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise e [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] nwinfo = self.network_api.allocate_for_instance( [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] created_port_ids = self._update_ports_for_instance( [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] with excutils.save_and_reraise_exception(): [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] self.force_reraise() [ 781.014490] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise self.value [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] updated_port = self._update_port( [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] _ensure_no_port_binding_failure(port) [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] raise exception.PortBindingFailed(port_id=port['id']) [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] nova.exception.PortBindingFailed: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. [ 781.014860] env[61947]: ERROR nova.compute.manager [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] [ 781.014860] env[61947]: DEBUG nova.compute.utils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.015421] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.468s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.017139] env[61947]: INFO nova.compute.claims [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.020164] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Build of instance 46dbb96b-dc8b-4fc9-bc54-181c5a308916 was re-scheduled: Binding failed for port 4f13b041-13bb-46c1-a43d-eaec3cd9a63e, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.020628] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.020998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquiring lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.020998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Acquired lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.021162] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.051078] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224216, 'name': Rename_Task, 'duration_secs': 1.150925} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.051078] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.051339] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-664831f7-e1bb-4c4b-9227-e6dd7ba54926 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.058458] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 781.058458] env[61947]: value = "task-1224220" [ 781.058458] env[61947]: _type = "Task" [ 781.058458] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.067708] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224220, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.103652] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.113300] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224218, 'name': ReconfigVM_Task, 'duration_secs': 0.261241} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.113668] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.114804] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ad2f842-3692-4663-89d7-a633af33e07d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.122010] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 781.122010] env[61947]: value = "task-1224221" [ 781.122010] env[61947]: _type = "Task" [ 781.122010] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.132242] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224221, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.134449] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.134845] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.134967] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.135166] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.135312] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.135455] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.136029] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.136029] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.136029] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.136295] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.136425] env[61947]: DEBUG nova.virt.hardware [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.137149] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcafb7f9-502b-4d92-86a3-64b264e72da7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.144952] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf01cf5b-b15c-4f74-b8dd-90f4e2671f26 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.200803] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442177} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.201083] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] d642bdb3-8e98-4667-9af7-71dae739e6b1/d642bdb3-8e98-4667-9af7-71dae739e6b1.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 781.201308] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.201562] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd99d1a4-2eb7-4231-ba52-bb6ff8dd6ebb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.208152] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 781.208152] env[61947]: value = "task-1224222" [ 781.208152] env[61947]: _type = "Task" [ 781.208152] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.215581] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.483753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Releasing lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.484112] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Instance network_info: |[{"id": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "address": "fa:16:3e:12:ba:7d", "network": {"id": "fdb2f6d6-01b5-4ea3-8b54-6ddb4f4f366d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1736584182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8aa679df5f264a438984f000b16f7d82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa67cfa09-3b", "ovs_interfaceid": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 781.484419] env[61947]: DEBUG oslo_concurrency.lockutils [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] Acquired lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.484602] env[61947]: DEBUG nova.network.neutron [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Refreshing network info cache for port a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.485837] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:ba:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a67cfa09-3b87-4186-8afb-19040cd3dcea', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.497696] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Creating folder: Project (8aa679df5f264a438984f000b16f7d82). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.502106] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10a45e4b-e32a-4636-96bb-1ff9bc40efb5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.513967] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Created folder: Project (8aa679df5f264a438984f000b16f7d82) in parent group-v264556. [ 781.513967] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Creating folder: Instances. Parent ref: group-v264583. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.514168] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3fde042-fb7d-47a3-8a0a-f6ac1d56dd4c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.528602] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Created folder: Instances in parent group-v264583. [ 781.528822] env[61947]: DEBUG oslo.service.loopingcall [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.529865] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.529865] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff3d81b3-10d3-4dcb-8cb9-2b2afaf13b68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.559820] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.559820] env[61947]: value = "task-1224225" [ 781.559820] env[61947]: _type = "Task" [ 781.559820] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.567410] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.573753] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224225, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.577612] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224220, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.632841] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224221, 'name': Rename_Task, 'duration_secs': 0.144392} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.633206] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.633494] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-365ab914-5e71-45a2-a4b8-59d28a1e1ba6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.640530] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 781.640530] env[61947]: value = "task-1224226" [ 781.640530] env[61947]: _type = "Task" [ 781.640530] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.648292] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224226, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.717956] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065742} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.720605] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.721632] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38479b4f-c7ff-4497-93a5-af7f98771625 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.752740] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] d642bdb3-8e98-4667-9af7-71dae739e6b1/d642bdb3-8e98-4667-9af7-71dae739e6b1.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.753095] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd699098-731b-468c-8d9b-017ca7cb8c5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.775018] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 781.775018] env[61947]: value = "task-1224227" [ 781.775018] env[61947]: _type = "Task" [ 781.775018] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.784374] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224227, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.815095] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.873023] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Successfully updated port: c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.944640] env[61947]: DEBUG nova.network.neutron [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Updated VIF entry in instance network info cache for port a67cfa09-3b87-4186-8afb-19040cd3dcea. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 781.945048] env[61947]: DEBUG nova.network.neutron [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Updating instance_info_cache with network_info: [{"id": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "address": "fa:16:3e:12:ba:7d", "network": {"id": "fdb2f6d6-01b5-4ea3-8b54-6ddb4f4f366d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1736584182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8aa679df5f264a438984f000b16f7d82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa67cfa09-3b", "ovs_interfaceid": "a67cfa09-3b87-4186-8afb-19040cd3dcea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.071424] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224225, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.077321] env[61947]: DEBUG oslo_vmware.api [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224220, 'name': PowerOnVM_Task, 'duration_secs': 0.737433} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.077564] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.077788] env[61947]: INFO nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Took 8.31 seconds to spawn the instance on the hypervisor. [ 782.077927] env[61947]: DEBUG nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.078779] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d38cff-7483-406b-9539-23ddc0ab852e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.150379] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224226, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.288127] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224227, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.322033] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Releasing lock "refresh_cache-46dbb96b-dc8b-4fc9-bc54-181c5a308916" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.322033] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.322033] env[61947]: DEBUG nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.322033] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.346888] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.353077] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b2e335-46a0-4ce4-9e25-87655e4ad491 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.361272] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576b3451-ecf4-481f-9dc9-70da9b420524 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.395441] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.395628] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.395783] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.397539] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0a9c5b-97b0-413b-84d8-bba8ba6fdaa1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.406923] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc98de1d-1c7e-4545-a621-6da460d4a77e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.423325] env[61947]: DEBUG nova.compute.provider_tree [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.447801] env[61947]: DEBUG oslo_concurrency.lockutils [req-ac5a1b60-6023-4ac6-a8c9-d703303342f0 req-8caab135-b07a-4e02-83f4-45f8d6adf939 service nova] Releasing lock "refresh_cache-c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.508757] env[61947]: DEBUG nova.compute.manager [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Received event network-vif-plugged-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.509123] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Acquiring lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.509390] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.509565] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.509744] env[61947]: DEBUG nova.compute.manager [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] No waiting events found dispatching network-vif-plugged-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.509913] env[61947]: WARNING nova.compute.manager [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Received unexpected event network-vif-plugged-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 for instance with vm_state building and task_state spawning. [ 782.510130] env[61947]: DEBUG nova.compute.manager [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Received event network-changed-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.510338] env[61947]: DEBUG nova.compute.manager [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Refreshing instance network info cache due to event network-changed-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 782.510390] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Acquiring lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.568099] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224225, 'name': CreateVM_Task, 'duration_secs': 0.596455} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.568414] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.568938] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.569138] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.569417] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.569656] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7e9349d-b027-445f-8a2d-bfe22c561deb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.573823] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 782.573823] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5231b9bc-21b3-bdb1-2d32-1cfaa909e470" [ 782.573823] env[61947]: _type = "Task" [ 782.573823] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.581340] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5231b9bc-21b3-bdb1-2d32-1cfaa909e470, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.595788] env[61947]: INFO nova.compute.manager [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Took 29.10 seconds to build instance. [ 782.650757] env[61947]: DEBUG oslo_vmware.api [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224226, 'name': PowerOnVM_Task, 'duration_secs': 0.834013} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.651061] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.651265] env[61947]: DEBUG nova.compute.manager [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.651993] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11fd507-5b75-4899-946e-43195016a26c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.784945] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224227, 'name': ReconfigVM_Task, 'duration_secs': 0.691042} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.785250] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Reconfigured VM instance instance-0000002d to attach disk [datastore2] d642bdb3-8e98-4667-9af7-71dae739e6b1/d642bdb3-8e98-4667-9af7-71dae739e6b1.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.785855] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf7495aa-592c-49c8-9a83-725c5549f07a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.791899] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 782.791899] env[61947]: value = "task-1224228" [ 782.791899] env[61947]: _type = "Task" [ 782.791899] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.799132] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224228, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.858026] env[61947]: DEBUG nova.network.neutron [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.927230] env[61947]: DEBUG nova.scheduler.client.report [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.943681] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.084994] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5231b9bc-21b3-bdb1-2d32-1cfaa909e470, 'name': SearchDatastore_Task, 'duration_secs': 0.009075} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.085424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.085424] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.085546] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.085931] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.085931] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.086213] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec3ce68a-e102-4d77-8902-3a038fd884ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.094321] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.094511] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.096963] env[61947]: DEBUG nova.network.neutron [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Updating instance_info_cache with network_info: [{"id": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "address": "fa:16:3e:54:56:8a", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ea69ed-6d", "ovs_interfaceid": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.096963] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c79a354b-1e68-4ed2-abe6-8f9c589918b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.101994] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8fc6d489-5cc1-49d1-8bd0-a2a8e5b8b942 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.137s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.105219] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 783.105219] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522cedea-3004-e4e2-25b2-d40a2b731071" [ 783.105219] env[61947]: _type = "Task" [ 783.105219] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.113868] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522cedea-3004-e4e2-25b2-d40a2b731071, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.165568] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.301828] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224228, 'name': Rename_Task, 'duration_secs': 0.148784} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.302131] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 783.302372] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34a6f020-afeb-4842-901c-8bab65cbae63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.308471] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 783.308471] env[61947]: value = "task-1224229" [ 783.308471] env[61947]: _type = "Task" [ 783.308471] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.315984] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.359543] env[61947]: INFO nova.compute.manager [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] [instance: 46dbb96b-dc8b-4fc9-bc54-181c5a308916] Took 1.04 seconds to deallocate network for instance. [ 783.432927] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.434030] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.437186] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.962s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.438864] env[61947]: INFO nova.compute.claims [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.602735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.603166] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance network_info: |[{"id": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "address": "fa:16:3e:54:56:8a", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ea69ed-6d", "ovs_interfaceid": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 783.603541] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.606351] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Acquired lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.606351] env[61947]: DEBUG nova.network.neutron [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Refreshing network info cache for port c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 783.607366] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:56:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.615946] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating folder: Project (58b83ea896bd4654aaa9e81f635a6a55). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.619461] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-208311d0-8e6b-4be7-98e1-70015936a311 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.634372] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522cedea-3004-e4e2-25b2-d40a2b731071, 'name': SearchDatastore_Task, 'duration_secs': 0.009171} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.636552] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created folder: Project (58b83ea896bd4654aaa9e81f635a6a55) in parent group-v264556. [ 783.636552] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating folder: Instances. Parent ref: group-v264586. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.636666] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77cf14f5-44da-4adb-bfbf-0c066787f40a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.638878] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03fcc432-e47b-4511-a8a5-bc3707a06fa6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.643746] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 783.643746] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d9438f-7ae9-2494-16ba-3e24de8d60cd" [ 783.643746] env[61947]: _type = "Task" [ 783.643746] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.649762] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created folder: Instances in parent group-v264586. [ 783.649762] env[61947]: DEBUG oslo.service.loopingcall [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.650091] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.650590] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db54d7b5-69f0-4c4a-a8bf-bccf78b043b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.670284] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d9438f-7ae9-2494-16ba-3e24de8d60cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.679569] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.679569] env[61947]: value = "task-1224232" [ 783.679569] env[61947]: _type = "Task" [ 783.679569] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.688487] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224232, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.722529] env[61947]: INFO nova.compute.manager [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Rebuilding instance [ 783.769582] env[61947]: DEBUG nova.compute.manager [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.770511] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a1799a-2e13-4799-bd01-c03e8a27668e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.818104] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224229, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.867609] env[61947]: DEBUG nova.network.neutron [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Updated VIF entry in instance network info cache for port c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 783.867963] env[61947]: DEBUG nova.network.neutron [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Updating instance_info_cache with network_info: [{"id": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "address": "fa:16:3e:54:56:8a", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ea69ed-6d", "ovs_interfaceid": "c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.943171] env[61947]: DEBUG nova.compute.utils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.948118] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.948118] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 783.991618] env[61947]: DEBUG nova.policy [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfbb74d3b05e45a0a1c01d5e5c9f957f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c43bf9e1b64647aa8934dc29d79e6dca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 784.133389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.153966] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d9438f-7ae9-2494-16ba-3e24de8d60cd, 'name': SearchDatastore_Task, 'duration_secs': 0.012374} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.155364] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.155364] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] c8f37b6f-4f0c-43db-a1ea-3c45d3956f66/c8f37b6f-4f0c-43db-a1ea-3c45d3956f66.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.155364] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fdf25b3-aab4-40db-afda-9562feb6cae6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.161380] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 784.161380] env[61947]: value = "task-1224233" [ 784.161380] env[61947]: _type = "Task" [ 784.161380] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.169396] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.190697] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224232, 'name': CreateVM_Task, 'duration_secs': 0.28431} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.190866] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.191548] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.191742] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.192093] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.192333] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95a4dbb6-5bc9-40fb-b062-0885447cca1f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.196898] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 784.196898] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52234395-0632-6384-00ed-87caae21ceab" [ 784.196898] env[61947]: _type = "Task" [ 784.196898] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.204722] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52234395-0632-6384-00ed-87caae21ceab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.283247] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 784.284728] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2d1ae95-5d07-4967-ac17-4d71c55d0eb0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.291542] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 784.291542] env[61947]: value = "task-1224234" [ 784.291542] env[61947]: _type = "Task" [ 784.291542] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.301026] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.324330] env[61947]: DEBUG oslo_vmware.api [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224229, 'name': PowerOnVM_Task, 'duration_secs': 0.74332} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.324330] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 784.324330] env[61947]: INFO nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Took 8.08 seconds to spawn the instance on the hypervisor. [ 784.324330] env[61947]: DEBUG nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 784.325102] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a610f7f2-4496-4506-974e-9fce7b1e88a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.349585] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Successfully created port: ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.373118] env[61947]: DEBUG oslo_concurrency.lockutils [req-3796f478-7106-41dc-aa9a-718adf1671aa req-03e3aedb-30ea-43bc-a279-299e41444d34 service nova] Releasing lock "refresh_cache-9a54ca4f-e6ec-4413-b162-fca0cd824e00" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.407413] env[61947]: INFO nova.scheduler.client.report [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Deleted allocations for instance 46dbb96b-dc8b-4fc9-bc54-181c5a308916 [ 784.447728] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.671979] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484704} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.676194] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] c8f37b6f-4f0c-43db-a1ea-3c45d3956f66/c8f37b6f-4f0c-43db-a1ea-3c45d3956f66.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 784.676511] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.676998] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62244cd4-a745-4e4c-b22c-00dd0ab93cfb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.683394] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 784.683394] env[61947]: value = "task-1224235" [ 784.683394] env[61947]: _type = "Task" [ 784.683394] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.696805] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224235, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.709369] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52234395-0632-6384-00ed-87caae21ceab, 'name': SearchDatastore_Task, 'duration_secs': 0.009294} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.709851] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.712308] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.712308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.712308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.712308] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.712308] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c4458d2-352c-48f4-812a-7ac90cf39001 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.718741] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.718926] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 784.719654] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8791c685-9cb5-40f3-9242-e3b26ad73961 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.727100] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 784.727100] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bebb8f-a205-6154-f47a-91af4128647d" [ 784.727100] env[61947]: _type = "Task" [ 784.727100] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.736302] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bebb8f-a205-6154-f47a-91af4128647d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.801342] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224234, 'name': PowerOffVM_Task, 'duration_secs': 0.133629} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.803930] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 784.804163] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 784.806102] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e8e397-f4ef-4c8f-ac0d-3b833728c080 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.809633] env[61947]: DEBUG nova.compute.manager [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Received event network-changed {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.809959] env[61947]: DEBUG nova.compute.manager [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Refreshing instance network info cache due to event network-changed. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.810258] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] Acquiring lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.810484] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] Acquired lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.810635] env[61947]: DEBUG nova.network.neutron [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.816573] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 784.817613] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6acce970-e1f6-42d5-becc-857207269d1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.819732] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f72c50-e6a1-4611-ab5d-c1f3649d47ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.828335] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a57906-fd23-4b73-9ace-87a363240249 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.867540] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e394744-c2f5-4aaa-a4c0-cb20c75e03e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.870017] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 784.870232] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 784.870432] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Deleting the datastore file [datastore2] 607d29ad-2db6-4146-a0df-192f727e9d31 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.870904] env[61947]: INFO nova.compute.manager [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Took 29.43 seconds to build instance. [ 784.871734] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af06b186-9a2d-4577-8c28-f74df4c80336 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.879629] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba7a43d-8458-4201-b632-acdce7c36a70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.884030] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 784.884030] env[61947]: value = "task-1224237" [ 784.884030] env[61947]: _type = "Task" [ 784.884030] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.897551] env[61947]: DEBUG nova.compute.provider_tree [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.906287] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224237, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.919077] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3d26f477-3313-42fe-a1f9-b2b08f8d149a tempest-ServerRescueNegativeTestJSON-1434139419 tempest-ServerRescueNegativeTestJSON-1434139419-project-member] Lock "46dbb96b-dc8b-4fc9-bc54-181c5a308916" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.897s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.193861] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224235, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059851} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.194157] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.195771] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ce2814-0b28-4164-adf7-bf85c9406e14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.217135] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] c8f37b6f-4f0c-43db-a1ea-3c45d3956f66/c8f37b6f-4f0c-43db-a1ea-3c45d3956f66.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.217413] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f4565a9-ea11-42ea-9832-c33c74352413 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.241015] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bebb8f-a205-6154-f47a-91af4128647d, 'name': SearchDatastore_Task, 'duration_secs': 0.00854} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.242763] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 785.242763] env[61947]: value = "task-1224238" [ 785.242763] env[61947]: _type = "Task" [ 785.242763] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.242966] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94f1fbfe-9290-4749-ac98-7151d1930fed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.250821] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 785.250821] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524ef579-fe2e-0ca5-f8e0-afcbbcdca4c3" [ 785.250821] env[61947]: _type = "Task" [ 785.250821] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.254074] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224238, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.260986] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524ef579-fe2e-0ca5-f8e0-afcbbcdca4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.374294] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b08fc20f-ece4-4589-83a7-6bb63151f350 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.665s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.393903] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224237, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109634} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.394758] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.394961] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 785.395162] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 785.401779] env[61947]: DEBUG nova.scheduler.client.report [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.421953] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.459860] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.488309] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.488591] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.488756] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.488936] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.489094] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.489270] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.489506] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.489670] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.489834] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.490028] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.490230] env[61947]: DEBUG nova.virt.hardware [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.491073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03566d61-dbad-4cf1-a153-a97575809672 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.499128] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c98836-d0b3-4427-86f1-030c2078810c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.527790] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.528050] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.528284] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.528471] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.528634] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.530734] env[61947]: INFO nova.compute.manager [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Terminating instance [ 785.533825] env[61947]: DEBUG nova.compute.manager [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.533964] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 785.534764] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1402e219-7d2a-491e-8ffd-a53db7bf462c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.542452] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 785.542688] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-551e9198-df7a-4051-9830-e99448f8459e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.549444] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 785.549444] env[61947]: value = "task-1224239" [ 785.549444] env[61947]: _type = "Task" [ 785.549444] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.559339] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.560774] env[61947]: DEBUG nova.network.neutron [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updating instance_info_cache with network_info: [{"id": "224bbfc2-7511-45fe-95b2-caf904d6794b", "address": "fa:16:3e:0c:ff:29", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.30", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap224bbfc2-75", "ovs_interfaceid": "224bbfc2-7511-45fe-95b2-caf904d6794b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.755946] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224238, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.764753] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524ef579-fe2e-0ca5-f8e0-afcbbcdca4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.765015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.765278] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 9a54ca4f-e6ec-4413-b162-fca0cd824e00/9a54ca4f-e6ec-4413-b162-fca0cd824e00.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 785.765530] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06710348-58da-4c60-8e80-34f64a0ead29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.773479] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 785.773479] env[61947]: value = "task-1224240" [ 785.773479] env[61947]: _type = "Task" [ 785.773479] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.784291] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.878602] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.907918] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.908471] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.910944] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.295s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.911448] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.911631] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 785.911929] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.558s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.913772] env[61947]: INFO nova.compute.claims [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.919173] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242be319-3070-4bb9-bfe6-c570785e7750 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.931460] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc8f235-fce5-433e-a0f1-b637ddf6f485 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.955774] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f4c48b-00f5-4825-8a94-5a6dd601a622 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.960889] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.967520] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1ef02e-5e24-4223-b452-405ff060dbc8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.007660] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181484MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 786.007934] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.040490] env[61947]: DEBUG nova.compute.manager [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Received event network-vif-plugged-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.040811] env[61947]: DEBUG oslo_concurrency.lockutils [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] Acquiring lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.041147] env[61947]: DEBUG oslo_concurrency.lockutils [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.041466] env[61947]: DEBUG oslo_concurrency.lockutils [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.041754] env[61947]: DEBUG nova.compute.manager [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] No waiting events found dispatching network-vif-plugged-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.042416] env[61947]: WARNING nova.compute.manager [req-dd66fe6c-8f18-4e26-a774-9b4227a54c3f req-4b26b00a-284f-4977-b222-0e7d56320145 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Received unexpected event network-vif-plugged-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb for instance with vm_state building and task_state spawning. [ 786.062135] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224239, 'name': PowerOffVM_Task, 'duration_secs': 0.181615} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.062455] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 786.062664] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 786.063271] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4e6d2698-9424-441a-9cef-72105969bcbe tempest-ServerExternalEventsTest-813304690 tempest-ServerExternalEventsTest-813304690-project] Releasing lock "refresh_cache-d642bdb3-8e98-4667-9af7-71dae739e6b1" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.063642] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4e207bd-575a-4954-a154-00cfaaf40291 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.100539] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Successfully updated port: ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.122448] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 786.122608] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 786.122791] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Deleting the datastore file [datastore2] d642bdb3-8e98-4667-9af7-71dae739e6b1 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.123065] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77e1b0fb-2c47-430f-bde0-f9762b43284f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.131831] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for the task: (returnval){ [ 786.131831] env[61947]: value = "task-1224242" [ 786.131831] env[61947]: _type = "Task" [ 786.131831] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.140743] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.255178] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224238, 'name': ReconfigVM_Task, 'duration_secs': 0.668545} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.255470] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Reconfigured VM instance instance-0000002e to attach disk [datastore2] c8f37b6f-4f0c-43db-a1ea-3c45d3956f66/c8f37b6f-4f0c-43db-a1ea-3c45d3956f66.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.256079] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e9c7d56-8489-4391-9c31-b4751fb02696 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.261893] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 786.261893] env[61947]: value = "task-1224243" [ 786.261893] env[61947]: _type = "Task" [ 786.261893] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.269246] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224243, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.281799] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432662} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.282089] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 9a54ca4f-e6ec-4413-b162-fca0cd824e00/9a54ca4f-e6ec-4413-b162-fca0cd824e00.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 786.282311] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 786.282560] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efaac693-f8f3-4d2c-8f30-2ba2b8148034 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.289885] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 786.289885] env[61947]: value = "task-1224244" [ 786.289885] env[61947]: _type = "Task" [ 786.289885] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.303014] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224244, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.410554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.421726] env[61947]: DEBUG nova.compute.utils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.426978] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.426978] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.446282] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.446282] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.446282] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.446282] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.446857] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.446857] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.446857] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.446857] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.447645] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.448013] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.448383] env[61947]: DEBUG nova.virt.hardware [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.449600] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dc7106-7911-4e0b-8433-13224de3dd89 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.459461] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560b9213-ca05-4c08-9cf8-d4670f1bb0fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.475783] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.481746] env[61947]: DEBUG oslo.service.loopingcall [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.481994] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 786.482259] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56ae3ee4-6c74-4405-a887-7d9359565b91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.501962] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.501962] env[61947]: value = "task-1224245" [ 786.501962] env[61947]: _type = "Task" [ 786.501962] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.511016] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224245, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.520327] env[61947]: DEBUG nova.policy [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e72a3a5ba35943b6a374abaa3e774115', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '634b11b08e6d4c2eb88d7eabe6c8e729', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 786.604185] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.604185] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquired lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.604185] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.642181] env[61947]: DEBUG oslo_vmware.api [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Task: {'id': task-1224242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195626} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.642535] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.642783] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 786.643422] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.643422] env[61947]: INFO nova.compute.manager [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 786.643605] env[61947]: DEBUG oslo.service.loopingcall [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.643873] env[61947]: DEBUG nova.compute.manager [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.644034] env[61947]: DEBUG nova.network.neutron [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.771909] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224243, 'name': Rename_Task, 'duration_secs': 0.226499} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.772436] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.772737] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48122147-6f3e-45f3-8d30-b4587dc29cd8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.780314] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 786.780314] env[61947]: value = "task-1224246" [ 786.780314] env[61947]: _type = "Task" [ 786.780314] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.788143] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.797817] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224244, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075062} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.798176] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.799126] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f44f2e-e936-4505-a8fd-130a6694bc2e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.820529] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 9a54ca4f-e6ec-4413-b162-fca0cd824e00/9a54ca4f-e6ec-4413-b162-fca0cd824e00.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.820806] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d80deca4-0cc0-475c-ba99-3ab9c7cf0b81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.840509] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 786.840509] env[61947]: value = "task-1224247" [ 786.840509] env[61947]: _type = "Task" [ 786.840509] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.848824] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224247, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.934412] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.019792] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224245, 'name': CreateVM_Task, 'duration_secs': 0.277078} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.022733] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.026579] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.026862] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.027254] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.027568] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d66b9ae-9b00-4ba3-ab82-0806c0af178f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.033508] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 787.033508] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523f38bc-942e-176f-b7b7-bf33b78842b6" [ 787.033508] env[61947]: _type = "Task" [ 787.033508] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.042818] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523f38bc-942e-176f-b7b7-bf33b78842b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.199575] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.290420] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224246, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.315833] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb5f055-69e6-4b2b-98e0-85bda1c01263 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.325100] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f79e1ce-0d9f-449d-9466-9646a00f7a36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.360899] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656bb8b1-c343-416e-8d5d-1736f20fd423 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.368623] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224247, 'name': ReconfigVM_Task, 'duration_secs': 0.337066} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.370813] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 9a54ca4f-e6ec-4413-b162-fca0cd824e00/9a54ca4f-e6ec-4413-b162-fca0cd824e00.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.371465] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d135675-6b14-469c-874d-cf019128a82f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.373984] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17cd92f-a2a6-48ff-a86b-eed57966e596 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.390915] env[61947]: DEBUG nova.compute.provider_tree [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.397014] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 787.397014] env[61947]: value = "task-1224248" [ 787.397014] env[61947]: _type = "Task" [ 787.397014] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.405546] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224248, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.545134] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523f38bc-942e-176f-b7b7-bf33b78842b6, 'name': SearchDatastore_Task, 'duration_secs': 0.009651} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.545462] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.545723] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.545917] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.546076] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.546261] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.548801] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db1346aa-b254-45a0-a976-6b8c11762615 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.560181] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Successfully created port: 6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.567283] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.567488] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 787.568260] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0401eb4a-69a7-4162-b417-7417636e24c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.576814] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 787.576814] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526c05e2-0c18-22b6-ef18-038337f2fe4c" [ 787.576814] env[61947]: _type = "Task" [ 787.576814] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.584213] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526c05e2-0c18-22b6-ef18-038337f2fe4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.694720] env[61947]: DEBUG nova.network.neutron [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updating instance_info_cache with network_info: [{"id": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "address": "fa:16:3e:9f:1f:6a", "network": {"id": "a4c804a4-bcde-4734-8ed7-44fb2d7f8ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1910379320-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c43bf9e1b64647aa8934dc29d79e6dca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0e00b2f1-c70f-4b21-86eb-810643cc1680", "external-id": "nsx-vlan-transportzone-487", "segmentation_id": 487, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae0ffc38-7e", "ovs_interfaceid": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.793170] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224246, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.897078] env[61947]: DEBUG nova.scheduler.client.report [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.912367] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224248, 'name': Rename_Task, 'duration_secs': 0.152237} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.912847] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.913466] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0153e13-2b17-48e4-9e5c-66c86c5863d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.921348] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 787.921348] env[61947]: value = "task-1224249" [ 787.921348] env[61947]: _type = "Task" [ 787.921348] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.932511] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.947606] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.976581] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.976821] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.976969] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.977180] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.977321] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.977631] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.977721] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.977805] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.977965] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.978259] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.978355] env[61947]: DEBUG nova.virt.hardware [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.979167] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a67cc9-4030-4393-bb27-3d0bddc5dec4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.987365] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe9ecbf-419d-4806-a698-0b582d0fa43a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.075189] env[61947]: DEBUG nova.compute.manager [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Received event network-changed-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.075430] env[61947]: DEBUG nova.compute.manager [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Refreshing instance network info cache due to event network-changed-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 788.075833] env[61947]: DEBUG oslo_concurrency.lockutils [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] Acquiring lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.087793] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526c05e2-0c18-22b6-ef18-038337f2fe4c, 'name': SearchDatastore_Task, 'duration_secs': 0.016545} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.088692] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e303c24-c2d8-4c4a-a35f-5787cd3beff8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.094703] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 788.094703] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a863c0-136f-4177-c8da-aa9ecc91b5b9" [ 788.094703] env[61947]: _type = "Task" [ 788.094703] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.104795] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a863c0-136f-4177-c8da-aa9ecc91b5b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.164644] env[61947]: DEBUG nova.network.neutron [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.197817] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Releasing lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.198183] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Instance network_info: |[{"id": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "address": "fa:16:3e:9f:1f:6a", "network": {"id": "a4c804a4-bcde-4734-8ed7-44fb2d7f8ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1910379320-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c43bf9e1b64647aa8934dc29d79e6dca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0e00b2f1-c70f-4b21-86eb-810643cc1680", "external-id": "nsx-vlan-transportzone-487", "segmentation_id": 487, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae0ffc38-7e", "ovs_interfaceid": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 788.198608] env[61947]: DEBUG oslo_concurrency.lockutils [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] Acquired lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.198683] env[61947]: DEBUG nova.network.neutron [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Refreshing network info cache for port ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.199919] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:1f:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0e00b2f1-c70f-4b21-86eb-810643cc1680', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae0ffc38-7ea0-412a-9169-2a2a27dda3eb', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.212295] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Creating folder: Project (c43bf9e1b64647aa8934dc29d79e6dca). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 788.218780] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c86e061-a558-4098-8332-792b06dc3af8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.231873] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Created folder: Project (c43bf9e1b64647aa8934dc29d79e6dca) in parent group-v264556. [ 788.232258] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Creating folder: Instances. Parent ref: group-v264590. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 788.232541] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b81d120-cd41-49e2-8b98-7da1816e0c6f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.241669] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Created folder: Instances in parent group-v264590. [ 788.242007] env[61947]: DEBUG oslo.service.loopingcall [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.242111] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 788.242849] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b52d6be8-805c-4c5b-8523-70b2adb6f29e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.264965] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.264965] env[61947]: value = "task-1224252" [ 788.264965] env[61947]: _type = "Task" [ 788.264965] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.273799] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224252, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.294183] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224246, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.406922] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.407617] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.411022] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.946s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.433395] env[61947]: DEBUG oslo_vmware.api [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224249, 'name': PowerOnVM_Task, 'duration_secs': 0.455923} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.434462] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.435057] env[61947]: INFO nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Took 7.33 seconds to spawn the instance on the hypervisor. [ 788.435281] env[61947]: DEBUG nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.436248] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4299240-fc72-4d21-a243-7563c86796bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.564077] env[61947]: DEBUG nova.network.neutron [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updated VIF entry in instance network info cache for port ae0ffc38-7ea0-412a-9169-2a2a27dda3eb. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 788.564549] env[61947]: DEBUG nova.network.neutron [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updating instance_info_cache with network_info: [{"id": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "address": "fa:16:3e:9f:1f:6a", "network": {"id": "a4c804a4-bcde-4734-8ed7-44fb2d7f8ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1910379320-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c43bf9e1b64647aa8934dc29d79e6dca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0e00b2f1-c70f-4b21-86eb-810643cc1680", "external-id": "nsx-vlan-transportzone-487", "segmentation_id": 487, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae0ffc38-7e", "ovs_interfaceid": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.606204] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a863c0-136f-4177-c8da-aa9ecc91b5b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.606434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.606758] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.607041] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f204990-131f-44c1-b125-bcf0f49905e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.613728] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 788.613728] env[61947]: value = "task-1224253" [ 788.613728] env[61947]: _type = "Task" [ 788.613728] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.621922] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.667129] env[61947]: INFO nova.compute.manager [-] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Took 2.02 seconds to deallocate network for instance. [ 788.776368] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224252, 'name': CreateVM_Task, 'duration_secs': 0.385243} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.776547] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 788.777288] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.777453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.777797] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.778132] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edf83de1-d1e5-4b88-b49c-9038091f1da7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.783632] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 788.783632] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb7491-81e4-74fa-ffef-7b7254485bd2" [ 788.783632] env[61947]: _type = "Task" [ 788.783632] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.795497] env[61947]: DEBUG oslo_vmware.api [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224246, 'name': PowerOnVM_Task, 'duration_secs': 1.68372} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.798698] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.798925] env[61947]: INFO nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Took 10.11 seconds to spawn the instance on the hypervisor. [ 788.799132] env[61947]: DEBUG nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.799448] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb7491-81e4-74fa-ffef-7b7254485bd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.800520] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da15f546-1e0a-42fa-8dad-9e9f0817011b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.916685] env[61947]: DEBUG nova.compute.utils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.923980] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.923980] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.959628] env[61947]: INFO nova.compute.manager [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Took 29.75 seconds to build instance. [ 788.991439] env[61947]: DEBUG nova.policy [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb456b8eab724ec7ba6245bf8d76104b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3a0a834fefd45198f88d7cd28621c5d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 789.068915] env[61947]: DEBUG oslo_concurrency.lockutils [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] Releasing lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.068915] env[61947]: DEBUG nova.compute.manager [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Received event network-vif-deleted-224bbfc2-7511-45fe-95b2-caf904d6794b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 789.068915] env[61947]: INFO nova.compute.manager [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Neutron deleted interface 224bbfc2-7511-45fe-95b2-caf904d6794b; detaching it from the instance and deleting it from the info cache [ 789.069119] env[61947]: DEBUG nova.network.neutron [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.130208] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44811} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.130571] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.130874] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.131263] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39a9c730-327b-4a16-8cf8-bd724d663503 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.141038] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 789.141038] env[61947]: value = "task-1224254" [ 789.141038] env[61947]: _type = "Task" [ 789.141038] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.150911] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224254, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.173989] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.303048] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb7491-81e4-74fa-ffef-7b7254485bd2, 'name': SearchDatastore_Task, 'duration_secs': 0.057441} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.304992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.305346] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.305914] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.306150] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.306460] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.307502] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b933fa4-e104-454a-bc78-9714ef44dd55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.314708] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baf60658-a9fc-4df7-bcdd-a65a23198c9f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.325702] env[61947]: INFO nova.compute.manager [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Took 31.98 seconds to build instance. [ 789.328792] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19a2104-5af7-4b46-9ceb-2074afbfb3df {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.333534] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.333534] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 789.334221] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aa0bfc2-9d62-4424-86c7-2522f6a62045 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.342929] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 789.342929] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52be0656-9a62-1f43-1d75-c8b5de284f35" [ 789.342929] env[61947]: _type = "Task" [ 789.342929] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.376791] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa23860-662c-4c56-9d50-4e220ea992c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.385279] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52be0656-9a62-1f43-1d75-c8b5de284f35, 'name': SearchDatastore_Task, 'duration_secs': 0.009795} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.389626] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32fcdef7-e026-4eff-ae75-38395ba3bcdc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.391344] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e87a60-2269-4f78-b00d-667f4c780419 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.397872] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 789.397872] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52935fb0-5efe-bf8f-4b60-5fc05b93c68a" [ 789.397872] env[61947]: _type = "Task" [ 789.397872] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.406393] env[61947]: DEBUG nova.compute.provider_tree [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.416255] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52935fb0-5efe-bf8f-4b60-5fc05b93c68a, 'name': SearchDatastore_Task, 'duration_secs': 0.009142} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.417075] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.417338] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b7d84a56-edb6-4d1e-b7e9-294078be79d8/b7d84a56-edb6-4d1e-b7e9-294078be79d8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.417590] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9112427f-05ab-4bdf-91d4-c73c45c54340 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.423038] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.427187] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 789.427187] env[61947]: value = "task-1224255" [ 789.427187] env[61947]: _type = "Task" [ 789.427187] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.437159] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.462486] env[61947]: DEBUG oslo_concurrency.lockutils [None req-db5b6c8b-2083-4fa3-b109-19476f86f4fa tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.017s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.573198] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97e867bf-3dac-4434-96fd-702ad16732d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.579083] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Successfully created port: 2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.588177] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189d982c-90a0-4d1d-9f72-13110d3ad9b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.609942] env[61947]: DEBUG nova.compute.manager [req-542ad3cf-b918-4ab5-96c5-a14144e2cb28 req-d243ee98-4bd1-437c-a64f-05992a59d856 service nova] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Detach interface failed, port_id=224bbfc2-7511-45fe-95b2-caf904d6794b, reason: Instance d642bdb3-8e98-4667-9af7-71dae739e6b1 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 789.650292] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224254, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.713272] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4d4959-6e9c-4c7d-8bf5-a229fa0b32ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.717410] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Successfully updated port: 6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.723007] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Suspending the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 789.724117] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-13f9e178-d9d3-4f7a-aebc-7373e84b1a6b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.734775] env[61947]: DEBUG oslo_vmware.api [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 789.734775] env[61947]: value = "task-1224256" [ 789.734775] env[61947]: _type = "Task" [ 789.734775] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.742850] env[61947]: DEBUG oslo_vmware.api [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224256, 'name': SuspendVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.835019] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6097341d-502b-46d0-9459-708d54397c64 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.320s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.915032] env[61947]: DEBUG nova.scheduler.client.report [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.946858] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492116} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.946858] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b7d84a56-edb6-4d1e-b7e9-294078be79d8/b7d84a56-edb6-4d1e-b7e9-294078be79d8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.946858] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.946858] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-513c9b0b-6227-452a-bd4d-eaeff680fdc1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.949677] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 789.949677] env[61947]: value = "task-1224257" [ 789.949677] env[61947]: _type = "Task" [ 789.949677] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.959338] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.967635] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.133015] env[61947]: DEBUG nova.compute.manager [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received event network-vif-plugged-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.133395] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Acquiring lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.133674] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.133929] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.134169] env[61947]: DEBUG nova.compute.manager [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] No waiting events found dispatching network-vif-plugged-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.134398] env[61947]: WARNING nova.compute.manager [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received unexpected event network-vif-plugged-6cde2663-2cd8-4148-bb88-3c01384dea78 for instance with vm_state building and task_state spawning. [ 790.134627] env[61947]: DEBUG nova.compute.manager [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.134830] env[61947]: DEBUG nova.compute.manager [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing instance network info cache due to event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.135047] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Acquiring lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.135317] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Acquired lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.135441] env[61947]: DEBUG nova.network.neutron [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing network info cache for port 6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.152407] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224254, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.220598] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.242651] env[61947]: DEBUG oslo_vmware.api [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224256, 'name': SuspendVM_Task} progress is 54%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.337747] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.421880] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.421880] env[61947]: ERROR nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Traceback (most recent call last): [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.driver.spawn(context, instance, image_meta, [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.421880] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] vm_ref = self.build_virtual_machine(instance, [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] for vif in network_info: [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return self._sync_wrapper(fn, *args, **kwargs) [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.wait() [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self[:] = self._gt.wait() [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return self._exit_event.wait() [ 790.422361] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] current.throw(*self._exc) [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] result = function(*args, **kwargs) [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] return func(*args, **kwargs) [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise e [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] nwinfo = self.network_api.allocate_for_instance( [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] created_port_ids = self._update_ports_for_instance( [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.422719] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] with excutils.save_and_reraise_exception(): [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] self.force_reraise() [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise self.value [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] updated_port = self._update_port( [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] _ensure_no_port_binding_failure(port) [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] raise exception.PortBindingFailed(port_id=port['id']) [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] nova.exception.PortBindingFailed: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. [ 790.423085] env[61947]: ERROR nova.compute.manager [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] [ 790.423421] env[61947]: DEBUG nova.compute.utils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 790.425015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.426s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.428566] env[61947]: INFO nova.compute.claims [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.429700] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Build of instance be1fdafa-3c4d-430c-99c5-202e6ccebf26 was re-scheduled: Binding failed for port 6170ba23-bb1e-42cd-8ccc-23ef1df38a96, please check neutron logs for more information. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 790.430164] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Unplugging VIFs for instance {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 790.430396] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.430564] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquired lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.430719] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.434082] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.460530] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.460784] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.460942] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.461139] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.461292] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.461443] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.461693] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.461840] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.462021] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.462183] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.462368] env[61947]: DEBUG nova.virt.hardware [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.463531] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd2c949-0d30-4748-aa9c-4bcb82a8f639 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.473243] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.230915} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.473243] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.473243] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d249d22-3449-4348-ae42-4b1afe3db570 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.480732] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd4c39f-d9d8-40dd-a2fe-2d4e49afa747 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.502902] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] b7d84a56-edb6-4d1e-b7e9-294078be79d8/b7d84a56-edb6-4d1e-b7e9-294078be79d8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.503756] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.504668] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f4b26f2-d800-4ff1-8800-5a903770eb8a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.531855] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 790.531855] env[61947]: value = "task-1224258" [ 790.531855] env[61947]: _type = "Task" [ 790.531855] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.539343] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224258, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.651361] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224254, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.304242} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.651608] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.652513] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b4348b-36a9-4c4d-84e4-38d6ee61492c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.674542] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.675559] env[61947]: DEBUG nova.network.neutron [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.677323] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f3d9e4d-4257-41eb-a6d5-45c790a1e7fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.697168] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 790.697168] env[61947]: value = "task-1224259" [ 790.697168] env[61947]: _type = "Task" [ 790.697168] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.709662] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224259, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.743542] env[61947]: DEBUG oslo_vmware.api [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224256, 'name': SuspendVM_Task, 'duration_secs': 0.675241} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.743928] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Suspended the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 790.744031] env[61947]: DEBUG nova.compute.manager [None req-a132bde6-3eb2-4ae3-898f-f6adedb70387 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.744815] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74c6158-95d6-4bed-b92e-14425b25a72e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.832734] env[61947]: DEBUG nova.network.neutron [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.862863] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.960998] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.041917] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224258, 'name': ReconfigVM_Task, 'duration_secs': 0.262092} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.042378] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Reconfigured VM instance instance-00000030 to attach disk [datastore1] b7d84a56-edb6-4d1e-b7e9-294078be79d8/b7d84a56-edb6-4d1e-b7e9-294078be79d8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.043053] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1bf2598-87c9-4b85-b01e-6915e8a7c0ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.049516] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 791.049516] env[61947]: value = "task-1224260" [ 791.049516] env[61947]: _type = "Task" [ 791.049516] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.057952] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224260, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.059912] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.207746] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224259, 'name': ReconfigVM_Task, 'duration_secs': 0.328916} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.208130] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 607d29ad-2db6-4146-a0df-192f727e9d31/607d29ad-2db6-4146-a0df-192f727e9d31.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.208808] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3681c47e-212d-4273-8643-96eac980412e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.215068] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 791.215068] env[61947]: value = "task-1224261" [ 791.215068] env[61947]: _type = "Task" [ 791.215068] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.223463] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224261, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.289234] env[61947]: DEBUG nova.compute.manager [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Received event network-vif-plugged-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 791.289876] env[61947]: DEBUG oslo_concurrency.lockutils [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] Acquiring lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.290124] env[61947]: DEBUG oslo_concurrency.lockutils [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.290302] env[61947]: DEBUG oslo_concurrency.lockutils [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.290512] env[61947]: DEBUG nova.compute.manager [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] No waiting events found dispatching network-vif-plugged-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.290624] env[61947]: WARNING nova.compute.manager [req-2b2ae129-15ff-441c-b79c-7d6d2edde72c req-081601d9-45c2-407d-8791-44334aedaf0b service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Received unexpected event network-vif-plugged-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 for instance with vm_state building and task_state spawning. [ 791.335287] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb5efc39-eb1a-445e-a2cc-35c1b6489cd5 req-fc3abccb-4b68-4b38-a6f7-6b64ce0b6674 service nova] Releasing lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.335905] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.336231] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.559993] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224260, 'name': Rename_Task, 'duration_secs': 0.141158} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.562338] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.563033] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d89ed54-738e-41e2-bf9f-8b071ba6d8c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.565769] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Releasing lock "refresh_cache-be1fdafa-3c4d-430c-99c5-202e6ccebf26" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.566241] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61947) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 791.566241] env[61947]: DEBUG nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.566384] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.568874] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 791.568874] env[61947]: value = "task-1224262" [ 791.568874] env[61947]: _type = "Task" [ 791.568874] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.578685] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224262, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.595008] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.725882] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224261, 'name': Rename_Task, 'duration_secs': 0.133052} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.726515] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.727266] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7a84da-2bf4-4b74-8455-530bb949c1fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.730101] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc08317a-6858-4557-9cf0-8a16f76712ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.736035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.736275] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.736493] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.736633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.736791] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.739905] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Successfully updated port: 2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.744030] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff48789-be01-4891-883f-f3c3af4ba9d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.747811] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Waiting for the task: (returnval){ [ 791.747811] env[61947]: value = "task-1224263" [ 791.747811] env[61947]: _type = "Task" [ 791.747811] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.748624] env[61947]: INFO nova.compute.manager [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Terminating instance [ 791.777772] env[61947]: DEBUG nova.compute.manager [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.778009] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 791.779598] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a725a12f-6e9e-474a-b291-a869354225e7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.783073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5d2a94-1809-4b5b-a20f-b5c4d7014d52 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.789071] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224263, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.794474] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 791.795639] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a724a1-bd4a-4d01-aeb6-760edbf4a5d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.799173] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6603aa6e-1adb-4042-86fc-9d3d20cfeede {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.811124] env[61947]: DEBUG nova.compute.provider_tree [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 791.813373] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 791.813373] env[61947]: value = "task-1224264" [ 791.813373] env[61947]: _type = "Task" [ 791.813373] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.821713] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.872469] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.029846] env[61947]: DEBUG nova.network.neutron [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [{"id": "6cde2663-2cd8-4148-bb88-3c01384dea78", "address": "fa:16:3e:73:9d:a0", "network": {"id": "9f7fc41f-6e00-4f25-b860-8c1443ca3c9d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-875635745-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634b11b08e6d4c2eb88d7eabe6c8e729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cde2663-2c", "ovs_interfaceid": "6cde2663-2cd8-4148-bb88-3c01384dea78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.079372] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224262, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.101289] env[61947]: DEBUG nova.network.neutron [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.224761] env[61947]: DEBUG nova.compute.manager [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Received event network-changed-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.225021] env[61947]: DEBUG nova.compute.manager [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Refreshing instance network info cache due to event network-changed-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 792.225255] env[61947]: DEBUG oslo_concurrency.lockutils [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] Acquiring lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.225400] env[61947]: DEBUG oslo_concurrency.lockutils [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] Acquired lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.225557] env[61947]: DEBUG nova.network.neutron [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Refreshing network info cache for port 2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.243423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.259511] env[61947]: DEBUG oslo_vmware.api [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Task: {'id': task-1224263, 'name': PowerOnVM_Task, 'duration_secs': 0.508657} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.260590] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 792.260590] env[61947]: DEBUG nova.compute.manager [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.261042] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b739f460-3df4-487b-b98d-d85d3323f031 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.325416] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224264, 'name': PowerOffVM_Task, 'duration_secs': 0.196077} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.325686] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.325850] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.326108] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49b1a140-b6c7-41a4-b2b6-5371548fbf33 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.333012] env[61947]: ERROR nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [req-2affd2bd-5d3e-4d27-a73d-408567162169] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2affd2bd-5d3e-4d27-a73d-408567162169"}]} [ 792.352579] env[61947]: DEBUG nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 792.365910] env[61947]: DEBUG nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 792.368281] env[61947]: DEBUG nova.compute.provider_tree [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.379053] env[61947]: DEBUG nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 792.399627] env[61947]: DEBUG nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 792.403189] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 792.403293] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 792.403452] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Deleting the datastore file [datastore2] c8f37b6f-4f0c-43db-a1ea-3c45d3956f66 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.403708] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bcd0a55d-dcd0-42d3-932b-b6b0ba9dde6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.409648] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for the task: (returnval){ [ 792.409648] env[61947]: value = "task-1224266" [ 792.409648] env[61947]: _type = "Task" [ 792.409648] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.418122] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.532405] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.532680] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance network_info: |[{"id": "6cde2663-2cd8-4148-bb88-3c01384dea78", "address": "fa:16:3e:73:9d:a0", "network": {"id": "9f7fc41f-6e00-4f25-b860-8c1443ca3c9d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-875635745-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634b11b08e6d4c2eb88d7eabe6c8e729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cde2663-2c", "ovs_interfaceid": "6cde2663-2cd8-4148-bb88-3c01384dea78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.533048] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:9d:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cde2663-2cd8-4148-bb88-3c01384dea78', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.540586] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Creating folder: Project (634b11b08e6d4c2eb88d7eabe6c8e729). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.540859] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-874a2d62-1269-4d0a-b9a8-d65a74d595c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.557534] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Created folder: Project (634b11b08e6d4c2eb88d7eabe6c8e729) in parent group-v264556. [ 792.557706] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Creating folder: Instances. Parent ref: group-v264593. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.558787] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0bcd001-63b5-4db0-9be4-4013f4a2505d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.567275] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Created folder: Instances in parent group-v264593. [ 792.567577] env[61947]: DEBUG oslo.service.loopingcall [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.569903] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.570322] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-981f6521-fe85-46e5-913d-fb19d0faadad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.597143] env[61947]: DEBUG oslo_vmware.api [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224262, 'name': PowerOnVM_Task, 'duration_secs': 0.549674} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.598348] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 792.598575] env[61947]: INFO nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Took 7.14 seconds to spawn the instance on the hypervisor. [ 792.598830] env[61947]: DEBUG nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.598983] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.598983] env[61947]: value = "task-1224269" [ 792.598983] env[61947]: _type = "Task" [ 792.598983] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.601645] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b09607c-89aa-44cf-b72c-098144934189 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.604458] env[61947]: INFO nova.compute.manager [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: be1fdafa-3c4d-430c-99c5-202e6ccebf26] Took 1.04 seconds to deallocate network for instance. [ 792.615579] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224269, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.689983] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9fbbba-fdcf-45c1-8a6a-a479168cc7ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.697470] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8effb2-44ee-4157-88d1-e7647d6167d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.727427] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c099282c-443c-466f-91c5-3c66d872a02f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.736494] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c5b174-70b2-478d-bce9-00e38dbc4990 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.749758] env[61947]: DEBUG nova.compute.provider_tree [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.769196] env[61947]: DEBUG nova.network.neutron [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.777249] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.855710] env[61947]: DEBUG nova.network.neutron [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.919299] env[61947]: DEBUG oslo_vmware.api [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Task: {'id': task-1224266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.394463} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.919616] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 792.919818] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 792.920039] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 792.920190] env[61947]: INFO nova.compute.manager [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Took 1.14 seconds to destroy the instance on the hypervisor. [ 792.920490] env[61947]: DEBUG oslo.service.loopingcall [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.920733] env[61947]: DEBUG nova.compute.manager [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.920889] env[61947]: DEBUG nova.network.neutron [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.060183] env[61947]: DEBUG nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.061108] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e65996c-fe7b-4bdc-bcea-e625f87cc19f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.117355] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224269, 'name': CreateVM_Task, 'duration_secs': 0.386812} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.117355] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.117606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.117726] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.117958] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.118268] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7b57eca-a30e-4c5f-add3-c66772a94e25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.126469] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 793.126469] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527f4e2d-11bd-6472-a0c9-519c95c304af" [ 793.126469] env[61947]: _type = "Task" [ 793.126469] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.133189] env[61947]: INFO nova.compute.manager [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Took 29.61 seconds to build instance. [ 793.137860] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527f4e2d-11bd-6472-a0c9-519c95c304af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.284451] env[61947]: DEBUG nova.scheduler.client.report [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 70 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 793.284659] env[61947]: DEBUG nova.compute.provider_tree [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 70 to 71 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 793.285317] env[61947]: DEBUG nova.compute.provider_tree [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.358837] env[61947]: DEBUG oslo_concurrency.lockutils [req-d21d568d-d453-43f8-a200-d123888bf104 req-6b3f84c7-0094-4445-9eaf-3c3152aaeb40 service nova] Releasing lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.359286] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquired lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.359458] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.573139] env[61947]: INFO nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] instance snapshotting [ 793.573437] env[61947]: WARNING nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 793.579028] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d8aa03-6b06-40aa-910f-0edff31975ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.603169] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cd1a26-6aca-41c0-8ccd-8548cb38a8c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.635026] env[61947]: DEBUG oslo_concurrency.lockutils [None req-91f5c053-0ade-4b82-99ac-815a9c4e603a tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.750s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.640872] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527f4e2d-11bd-6472-a0c9-519c95c304af, 'name': SearchDatastore_Task, 'duration_secs': 0.018923} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.641205] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.641439] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.641663] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.641894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.641970] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.642838] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97dde9e7-8f84-416e-a14d-98693a093309 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.646967] env[61947]: INFO nova.scheduler.client.report [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Deleted allocations for instance be1fdafa-3c4d-430c-99c5-202e6ccebf26 [ 793.653688] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.653864] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.655749] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50f70efb-7d88-4b0d-811f-44a3a9f5bed1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.661559] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 793.661559] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254d49c-90ee-a69d-0590-d474eff87ac5" [ 793.661559] env[61947]: _type = "Task" [ 793.661559] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.670336] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254d49c-90ee-a69d-0590-d474eff87ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.792027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.368s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.792487] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.797707] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.632s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.797707] env[61947]: DEBUG nova.objects.instance [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61947) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 793.876441] env[61947]: DEBUG nova.network.neutron [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.901649] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.050501] env[61947]: DEBUG nova.network.neutron [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Updating instance_info_cache with network_info: [{"id": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "address": "fa:16:3e:81:16:cc", "network": {"id": "cf403b17-d3cc-4a45-91a2-a32571c485eb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1002914311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3a0a834fefd45198f88d7cd28621c5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a7b991a-9f", "ovs_interfaceid": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.103102] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "607d29ad-2db6-4146-a0df-192f727e9d31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.103365] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.103572] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "607d29ad-2db6-4146-a0df-192f727e9d31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.103764] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.103934] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.105969] env[61947]: INFO nova.compute.manager [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Terminating instance [ 794.107614] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.107764] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquired lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.107925] env[61947]: DEBUG nova.network.neutron [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.113986] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 794.114253] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-051fba2a-d954-41ff-a563-3c7d51d4a98c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.121746] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 794.121746] env[61947]: value = "task-1224270" [ 794.121746] env[61947]: _type = "Task" [ 794.121746] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.130523] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224270, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.143547] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.161026] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb6d1f6-2f2a-424f-9880-094a28c8bc58 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "be1fdafa-3c4d-430c-99c5-202e6ccebf26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.371s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.172496] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254d49c-90ee-a69d-0590-d474eff87ac5, 'name': SearchDatastore_Task, 'duration_secs': 0.009565} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.173292] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-416a8385-b34e-4d3a-b1b0-b13906e2f62a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.179172] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 794.179172] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f42fb-04a5-fe82-91e2-3390cb17724a" [ 794.179172] env[61947]: _type = "Task" [ 794.179172] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.187380] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f42fb-04a5-fe82-91e2-3390cb17724a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.251481] env[61947]: DEBUG nova.compute.manager [req-f15707b1-bcb4-4c2e-a68b-3de6d902aae7 req-5f131918-2214-4dc5-9d4e-6219ac2b2650 service nova] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Received event network-vif-deleted-a67cfa09-3b87-4186-8afb-19040cd3dcea {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.303322] env[61947]: DEBUG nova.compute.utils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.308166] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.308166] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.378097] env[61947]: INFO nova.compute.manager [-] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Took 1.46 seconds to deallocate network for instance. [ 794.385910] env[61947]: DEBUG nova.policy [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '296019cae8fe471aa727350d42542a9d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9829fe01401345e982276332b9b901cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 794.552931] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Releasing lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.558018] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance network_info: |[{"id": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "address": "fa:16:3e:81:16:cc", "network": {"id": "cf403b17-d3cc-4a45-91a2-a32571c485eb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1002914311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3a0a834fefd45198f88d7cd28621c5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a7b991a-9f", "ovs_interfaceid": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.558179] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:16:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a7b991a-9fca-4ed2-8096-6317d8bdc7f4', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.563346] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Creating folder: Project (d3a0a834fefd45198f88d7cd28621c5d). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 794.566496] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85b9a885-d0fa-438e-bfb9-90d392ea621b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.569052] env[61947]: DEBUG nova.compute.manager [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Received event network-changed-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.569145] env[61947]: DEBUG nova.compute.manager [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Refreshing instance network info cache due to event network-changed-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.569424] env[61947]: DEBUG oslo_concurrency.lockutils [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] Acquiring lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.569506] env[61947]: DEBUG oslo_concurrency.lockutils [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] Acquired lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.569663] env[61947]: DEBUG nova.network.neutron [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Refreshing network info cache for port ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.586968] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Created folder: Project (d3a0a834fefd45198f88d7cd28621c5d) in parent group-v264556. [ 794.587427] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Creating folder: Instances. Parent ref: group-v264596. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 794.587517] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a6d81ab-5fe0-4891-a498-c32cdff8047d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.596547] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Created folder: Instances in parent group-v264596. [ 794.596804] env[61947]: DEBUG oslo.service.loopingcall [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.597019] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 794.597260] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9a7e8f6-7887-4f41-a4ae-f46aa7488791 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.621274] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.621274] env[61947]: value = "task-1224273" [ 794.621274] env[61947]: _type = "Task" [ 794.621274] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.633587] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224273, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.637992] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224270, 'name': CreateSnapshot_Task, 'duration_secs': 0.50111} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.637992] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 794.637992] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75f51d6-319e-4c4f-a86c-9d7439cf52fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.641225] env[61947]: DEBUG nova.network.neutron [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.664201] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.671923] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.692456] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f42fb-04a5-fe82-91e2-3390cb17724a, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.692909] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.693210] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 2fc54703-5438-4c2a-b1e7-77431ce5177d/2fc54703-5438-4c2a-b1e7-77431ce5177d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 794.695360] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fed355a1-9174-46f9-849b-e815d3b775a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.702084] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 794.702084] env[61947]: value = "task-1224274" [ 794.702084] env[61947]: _type = "Task" [ 794.702084] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.715866] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224274, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.745602] env[61947]: DEBUG nova.network.neutron [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.771645] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Successfully created port: a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.808926] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.813329] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b68b6a54-1c38-44fa-b57c-8a168ee08d01 tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.818022] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.681s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.818022] env[61947]: INFO nova.compute.claims [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.888357] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.037401] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "683d40c9-b738-486a-806a-7e895637ee85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.037691] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.037907] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "683d40c9-b738-486a-806a-7e895637ee85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.038175] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.038371] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.044346] env[61947]: INFO nova.compute.manager [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Terminating instance [ 795.046716] env[61947]: DEBUG nova.compute.manager [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.046932] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.048036] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f951e92-e01a-46c1-9e53-b31f1eabbf76 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.056778] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.057090] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a887d7b-9786-43d4-80e0-ff8dbaa717a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.064166] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 795.064166] env[61947]: value = "task-1224275" [ 795.064166] env[61947]: _type = "Task" [ 795.064166] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.073102] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.132454] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224273, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.164660] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 795.167466] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c22c28fa-de8f-4bd8-8303-1c2e7fabccb9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.181438] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 795.181438] env[61947]: value = "task-1224276" [ 795.181438] env[61947]: _type = "Task" [ 795.181438] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.190597] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224276, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.196815] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.215697] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224274, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.249246] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Releasing lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.249700] env[61947]: DEBUG nova.compute.manager [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.249898] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.250780] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ce53f7-ad35-442b-8944-60132a216d71 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.257926] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.258213] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a7fb758-7b75-4f6a-b80c-52b65993c97a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.263899] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 795.263899] env[61947]: value = "task-1224277" [ 795.263899] env[61947]: _type = "Task" [ 795.263899] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.272512] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.401873] env[61947]: DEBUG nova.network.neutron [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updated VIF entry in instance network info cache for port ae0ffc38-7ea0-412a-9169-2a2a27dda3eb. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.402341] env[61947]: DEBUG nova.network.neutron [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updating instance_info_cache with network_info: [{"id": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "address": "fa:16:3e:9f:1f:6a", "network": {"id": "a4c804a4-bcde-4734-8ed7-44fb2d7f8ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1910379320-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c43bf9e1b64647aa8934dc29d79e6dca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0e00b2f1-c70f-4b21-86eb-810643cc1680", "external-id": "nsx-vlan-transportzone-487", "segmentation_id": 487, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae0ffc38-7e", "ovs_interfaceid": "ae0ffc38-7ea0-412a-9169-2a2a27dda3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.574775] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.631997] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224273, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.690920] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224276, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.714968] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224274, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52019} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.715599] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 2fc54703-5438-4c2a-b1e7-77431ce5177d/2fc54703-5438-4c2a-b1e7-77431ce5177d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 795.715599] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.715763] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3551d64-d651-4a2b-9999-aa134166e20e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.722320] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 795.722320] env[61947]: value = "task-1224278" [ 795.722320] env[61947]: _type = "Task" [ 795.722320] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.730186] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224278, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.774849] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224277, 'name': PowerOffVM_Task, 'duration_secs': 0.206301} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.775383] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 795.775652] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 795.776099] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c36d058d-0073-46d5-85f8-2e1ac3544ef6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.802449] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 795.802693] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 795.802875] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Deleting the datastore file [datastore1] 607d29ad-2db6-4146-a0df-192f727e9d31 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.803185] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-398b2d8f-2d19-4065-b3e1-6a692b83b217 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.810282] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for the task: (returnval){ [ 795.810282] env[61947]: value = "task-1224280" [ 795.810282] env[61947]: _type = "Task" [ 795.810282] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.817946] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224280, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.824008] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.850667] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.850994] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.851178] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.851368] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.851506] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.851648] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.851852] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.852026] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.852198] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.852362] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.852531] env[61947]: DEBUG nova.virt.hardware [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.853714] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de163c98-0658-4c12-9f76-fac193463f19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.864228] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fb00ac-f7cd-4b1e-8d36-7c574445c882 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.905381] env[61947]: DEBUG oslo_concurrency.lockutils [req-22dcc713-5281-4864-8ad0-98bb532957bc req-a0183bfe-d28f-4869-a3b4-edbb52db1e06 service nova] Releasing lock "refresh_cache-b7d84a56-edb6-4d1e-b7e9-294078be79d8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.079650] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224275, 'name': PowerOffVM_Task, 'duration_secs': 0.992338} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.079925] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 796.080117] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 796.080376] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2970d712-05f9-47d4-861e-1f2f746b22b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.129225] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bca924-1523-47e2-a3fa-060174771391 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.134987] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224273, 'name': CreateVM_Task, 'duration_secs': 1.313692} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.136302] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 796.136589] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 796.136775] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 796.136945] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Deleting the datastore file [datastore2] 683d40c9-b738-486a-806a-7e895637ee85 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.137606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.137763] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.138108] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.138334] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13fa1f99-dbfa-4761-a6c8-65463e50545c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.141531] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bc26a1a-1b58-4303-a021-3e8ec8bce351 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.144006] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9b1b84-6363-4c3f-b9f3-b03ff550d9b0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.148374] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for the task: (returnval){ [ 796.148374] env[61947]: value = "task-1224282" [ 796.148374] env[61947]: _type = "Task" [ 796.148374] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.176326] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 796.176326] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52830631-d7aa-c610-5755-3559485ee7e4" [ 796.176326] env[61947]: _type = "Task" [ 796.176326] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.177728] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645fc568-e8e2-4d49-b8a8-99a0e643a234 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.186022] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.192447] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abd8f06-8839-486c-bdbf-aec53a0a75ce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.201887] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224276, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.202086] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52830631-d7aa-c610-5755-3559485ee7e4, 'name': SearchDatastore_Task, 'duration_secs': 0.026424} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.202675] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.202916] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.203171] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.203321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.203495] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.203765] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1347bb02-b85e-4c8b-b4b5-a38a7f5eed00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.216411] env[61947]: DEBUG nova.compute.provider_tree [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.218582] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.218756] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 796.219993] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5be5510-1f63-422d-aad4-6c57ff90ceed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.229254] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 796.229254] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527a544a-3401-003e-b4c0-bb3a64a68833" [ 796.229254] env[61947]: _type = "Task" [ 796.229254] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.232726] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224278, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067277} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.235824] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.236570] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea4d9b7-aa93-4532-aba2-565cd1bcae56 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.245283] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527a544a-3401-003e-b4c0-bb3a64a68833, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.262869] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 2fc54703-5438-4c2a-b1e7-77431ce5177d/2fc54703-5438-4c2a-b1e7-77431ce5177d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.263197] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25aeea1a-6f1b-4beb-9828-d159938a9268 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.282915] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 796.282915] env[61947]: value = "task-1224283" [ 796.282915] env[61947]: _type = "Task" [ 796.282915] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.292076] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224283, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.319519] env[61947]: DEBUG oslo_vmware.api [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Task: {'id': task-1224280, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235769} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.319880] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.320098] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 796.320189] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 796.320343] env[61947]: INFO nova.compute.manager [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Took 1.07 seconds to destroy the instance on the hypervisor. [ 796.320758] env[61947]: DEBUG oslo.service.loopingcall [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.320969] env[61947]: DEBUG nova.compute.manager [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.321104] env[61947]: DEBUG nova.network.neutron [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.343726] env[61947]: DEBUG nova.network.neutron [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.452894] env[61947]: DEBUG nova.compute.manager [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Received event network-vif-plugged-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.453154] env[61947]: DEBUG oslo_concurrency.lockutils [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] Acquiring lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.453375] env[61947]: DEBUG oslo_concurrency.lockutils [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.453543] env[61947]: DEBUG oslo_concurrency.lockutils [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.453706] env[61947]: DEBUG nova.compute.manager [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] No waiting events found dispatching network-vif-plugged-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.453868] env[61947]: WARNING nova.compute.manager [req-45ff1bb5-de44-417e-8537-6632486b2cbe req-6b37966c-70a9-4e19-9cbf-215e0767e4c0 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Received unexpected event network-vif-plugged-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d for instance with vm_state building and task_state spawning. [ 796.568194] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Successfully updated port: a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.659263] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.695649] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224276, 'name': CloneVM_Task, 'duration_secs': 1.45095} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.695928] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Created linked-clone VM from snapshot [ 796.696740] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42ded3d-60b1-41e5-a1df-911871774276 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.705522] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Uploading image 48c8bce4-a273-4cde-9782-3b9274b22e19 {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 796.727651] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 796.727651] env[61947]: value = "vm-264600" [ 796.727651] env[61947]: _type = "VirtualMachine" [ 796.727651] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 796.728661] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4c97e1bb-b358-4fdc-b4e2-8f15d4be4cc6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.737705] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease: (returnval){ [ 796.737705] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282ef2e-c7f6-27ed-141d-d6b2a412d722" [ 796.737705] env[61947]: _type = "HttpNfcLease" [ 796.737705] env[61947]: } obtained for exporting VM: (result){ [ 796.737705] env[61947]: value = "vm-264600" [ 796.737705] env[61947]: _type = "VirtualMachine" [ 796.737705] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 796.738030] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the lease: (returnval){ [ 796.738030] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282ef2e-c7f6-27ed-141d-d6b2a412d722" [ 796.738030] env[61947]: _type = "HttpNfcLease" [ 796.738030] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 796.745209] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527a544a-3401-003e-b4c0-bb3a64a68833, 'name': SearchDatastore_Task, 'duration_secs': 0.028754} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.746395] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91df82a9-9374-4eec-9473-16a13ef7986c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.750096] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 796.750096] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282ef2e-c7f6-27ed-141d-d6b2a412d722" [ 796.750096] env[61947]: _type = "HttpNfcLease" [ 796.750096] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 796.753329] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 796.753329] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283050f-d8f7-b8ce-8614-5cdd264a275e" [ 796.753329] env[61947]: _type = "Task" [ 796.753329] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.753661] env[61947]: DEBUG nova.scheduler.client.report [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 71 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 796.753844] env[61947]: DEBUG nova.compute.provider_tree [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 71 to 72 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 796.754037] env[61947]: DEBUG nova.compute.provider_tree [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.765522] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283050f-d8f7-b8ce-8614-5cdd264a275e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.793007] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.846622] env[61947]: DEBUG nova.network.neutron [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.067790] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.068028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquired lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.068236] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.159639] env[61947]: DEBUG oslo_vmware.api [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Task: {'id': task-1224282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.536988} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.159881] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.160080] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.160261] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.160429] env[61947]: INFO nova.compute.manager [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Took 2.11 seconds to destroy the instance on the hypervisor. [ 797.160659] env[61947]: DEBUG oslo.service.loopingcall [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.160846] env[61947]: DEBUG nova.compute.manager [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.160939] env[61947]: DEBUG nova.network.neutron [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.246385] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 797.246385] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282ef2e-c7f6-27ed-141d-d6b2a412d722" [ 797.246385] env[61947]: _type = "HttpNfcLease" [ 797.246385] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 797.246680] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 797.246680] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282ef2e-c7f6-27ed-141d-d6b2a412d722" [ 797.246680] env[61947]: _type = "HttpNfcLease" [ 797.246680] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 797.247413] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe6a16b-aeb2-4ddc-88bf-82263cd13fa3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.257996] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 797.257996] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 797.316884] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.317517] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.321729] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.361s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.323184] env[61947]: INFO nova.compute.claims [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.334310] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283050f-d8f7-b8ce-8614-5cdd264a275e, 'name': SearchDatastore_Task, 'duration_secs': 0.013116} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.334661] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.334948] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6/7612f830-fed2-4dd7-ba8f-f792e1aa5ac6.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 797.335248] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3035bef6-23f9-4e60-9fb9-e37a54c6af9a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.340749] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224283, 'name': ReconfigVM_Task, 'duration_secs': 0.686004} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.341390] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 2fc54703-5438-4c2a-b1e7-77431ce5177d/2fc54703-5438-4c2a-b1e7-77431ce5177d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.342057] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b3b17b3-8d5d-4977-b587-89e34a4760aa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.346052] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 797.346052] env[61947]: value = "task-1224285" [ 797.346052] env[61947]: _type = "Task" [ 797.346052] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.349585] env[61947]: INFO nova.compute.manager [-] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Took 1.03 seconds to deallocate network for instance. [ 797.349931] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 797.349931] env[61947]: value = "task-1224286" [ 797.349931] env[61947]: _type = "Task" [ 797.349931] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.364224] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.364224] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c5de3a43-5dcf-4b32-b6fb-399d9f79f661 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.369548] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224286, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.604390] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.832744] env[61947]: DEBUG nova.compute.utils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.837486] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.841211] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.868900] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.874856] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224285, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.882281] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224286, 'name': Rename_Task, 'duration_secs': 0.137953} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.883360] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.883360] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53e27389-29eb-469b-8281-5e002d2fc0ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.893486] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 797.893486] env[61947]: value = "task-1224287" [ 797.893486] env[61947]: _type = "Task" [ 797.893486] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.901023] env[61947]: DEBUG nova.network.neutron [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Updating instance_info_cache with network_info: [{"id": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "address": "fa:16:3e:ac:bb:0e", "network": {"id": "468522c2-5b48-42f2-88cb-7a9f9cdae2e5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1441122954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9829fe01401345e982276332b9b901cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6de66aa-4d", "ovs_interfaceid": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.908407] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.927023] env[61947]: DEBUG nova.policy [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '382358439d2c4512b1450ed16bf277bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5273b44dfca848538176692ab2a19e3d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 797.954114] env[61947]: DEBUG nova.network.neutron [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.341398] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.363900] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592721} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.364501] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6/7612f830-fed2-4dd7-ba8f-f792e1aa5ac6.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 798.364868] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.365278] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd46822f-e04a-4a16-bea9-3095fa6614fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.376215] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 798.376215] env[61947]: value = "task-1224288" [ 798.376215] env[61947]: _type = "Task" [ 798.376215] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.392787] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224288, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.413043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Releasing lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.413043] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Instance network_info: |[{"id": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "address": "fa:16:3e:ac:bb:0e", "network": {"id": "468522c2-5b48-42f2-88cb-7a9f9cdae2e5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1441122954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9829fe01401345e982276332b9b901cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6de66aa-4d", "ovs_interfaceid": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.413441] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224287, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.415245] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:bb:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4020f51-6e46-4b73-a79e-9fe3fd51b917', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.423298] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Creating folder: Project (9829fe01401345e982276332b9b901cc). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.424162] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-280709e7-08e2-468f-a440-24631af33698 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.432783] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Successfully created port: 46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.438896] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Created folder: Project (9829fe01401345e982276332b9b901cc) in parent group-v264556. [ 798.439306] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Creating folder: Instances. Parent ref: group-v264601. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.439763] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eae8bcc9-c5d1-4296-bc71-704594300517 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.455361] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Created folder: Instances in parent group-v264601. [ 798.455611] env[61947]: DEBUG oslo.service.loopingcall [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.455998] env[61947]: INFO nova.compute.manager [-] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Took 1.29 seconds to deallocate network for instance. [ 798.456256] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.461507] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e94eaa6a-d194-49cd-a675-0232edafcd93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.489671] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.489671] env[61947]: value = "task-1224291" [ 798.489671] env[61947]: _type = "Task" [ 798.489671] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.499345] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224291, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.554577] env[61947]: DEBUG nova.compute.manager [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Received event network-changed-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.554986] env[61947]: DEBUG nova.compute.manager [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Refreshing instance network info cache due to event network-changed-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 798.555498] env[61947]: DEBUG oslo_concurrency.lockutils [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] Acquiring lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.555744] env[61947]: DEBUG oslo_concurrency.lockutils [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] Acquired lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.555884] env[61947]: DEBUG nova.network.neutron [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Refreshing network info cache for port a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.750847] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fb29c8-9b83-42dd-be54-e3bdbef632cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.758826] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca03c09-4700-40ca-8c86-fdb54d75d093 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.791270] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc01de54-86be-42a0-b646-633068971730 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.799190] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454e0310-59d7-4d40-8b6b-706d22382f7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.814659] env[61947]: DEBUG nova.compute.provider_tree [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.890177] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224288, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070587} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.890477] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.891264] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da71204-07c2-41fd-bcf9-f28096639da8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.915826] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6/7612f830-fed2-4dd7-ba8f-f792e1aa5ac6.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.916608] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d4cd91d-d161-413d-a58e-39b7a67b8c17 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.934642] env[61947]: DEBUG oslo_vmware.api [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224287, 'name': PowerOnVM_Task, 'duration_secs': 0.675708} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.935138] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.935369] env[61947]: INFO nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Took 10.99 seconds to spawn the instance on the hypervisor. [ 798.935558] env[61947]: DEBUG nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.936356] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4d0ca4-c460-49d5-8e23-deaa75872c0e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.939993] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 798.939993] env[61947]: value = "task-1224292" [ 798.939993] env[61947]: _type = "Task" [ 798.939993] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.951711] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224292, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.985357] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.999820] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224291, 'name': CreateVM_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.301287] env[61947]: DEBUG nova.network.neutron [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Updated VIF entry in instance network info cache for port a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.301764] env[61947]: DEBUG nova.network.neutron [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Updating instance_info_cache with network_info: [{"id": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "address": "fa:16:3e:ac:bb:0e", "network": {"id": "468522c2-5b48-42f2-88cb-7a9f9cdae2e5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1441122954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9829fe01401345e982276332b9b901cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6de66aa-4d", "ovs_interfaceid": "a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.317913] env[61947]: DEBUG nova.scheduler.client.report [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.349342] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.386702] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.387036] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.387121] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.387283] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.387432] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.387579] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.387786] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.387944] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.388152] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.388338] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.388514] env[61947]: DEBUG nova.virt.hardware [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.389451] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea875ccb-4887-4774-8c0e-081fb8c7a2a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.398027] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4261d0f2-3727-457e-bf23-eac722f2b64d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.461243] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224292, 'name': ReconfigVM_Task, 'duration_secs': 0.317108} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.461728] env[61947]: INFO nova.compute.manager [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Took 34.00 seconds to build instance. [ 799.462596] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6/7612f830-fed2-4dd7-ba8f-f792e1aa5ac6.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.464118] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d92c70bb-6f48-4060-a810-7f0d7e9b412b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.473051] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 799.473051] env[61947]: value = "task-1224293" [ 799.473051] env[61947]: _type = "Task" [ 799.473051] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.481160] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224293, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.500471] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224291, 'name': CreateVM_Task, 'duration_secs': 0.670036} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.500471] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 799.500824] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.501153] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.501347] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.501608] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69c99dd1-3d02-4724-a651-c1dcfee0ab8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.506351] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 799.506351] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522549fc-ded5-4049-8cb7-ba34eacc03a0" [ 799.506351] env[61947]: _type = "Task" [ 799.506351] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.514735] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522549fc-ded5-4049-8cb7-ba34eacc03a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.804805] env[61947]: DEBUG oslo_concurrency.lockutils [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] Releasing lock "refresh_cache-9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.805258] env[61947]: DEBUG nova.compute.manager [req-550c0169-8363-4874-a2ff-752c9f637901 req-3e71b938-9034-4e7f-b3a1-586f3d12e918 service nova] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Received event network-vif-deleted-0d86518d-eb0e-4c5b-996b-c88ec7987305 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.824485] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.825098] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.827877] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.820s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.964743] env[61947]: DEBUG oslo_concurrency.lockutils [None req-abebeee8-c99d-4b28-b866-48399f39e886 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.204s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.982994] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224293, 'name': Rename_Task, 'duration_secs': 0.136204} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.983323] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 799.983572] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f677ab3a-2cd4-4fb0-b9d5-0e4750aad9fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.990327] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 799.990327] env[61947]: value = "task-1224294" [ 799.990327] env[61947]: _type = "Task" [ 799.990327] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.998613] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.022099] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522549fc-ded5-4049-8cb7-ba34eacc03a0, 'name': SearchDatastore_Task, 'duration_secs': 0.016484} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.022433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.022691] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.022921] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.023105] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.023373] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.023610] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-952e39a1-cfd9-4445-9be7-153e4ffd9b42 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.026854] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Successfully updated port: 46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.035668] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.035911] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 800.036858] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9d7734a-ad00-4191-8612-04286de2ae53 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.043650] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 800.043650] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d5a3d9-1866-4e48-835a-79b69dc4460f" [ 800.043650] env[61947]: _type = "Task" [ 800.043650] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.053438] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d5a3d9-1866-4e48-835a-79b69dc4460f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.331544] env[61947]: DEBUG nova.compute.utils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.341952] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.341952] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.421850] env[61947]: DEBUG nova.policy [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc31d22cee2a4464937fc6d857dd28db', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1edda8c483e448c6a539df24264a14a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 800.467473] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.501790] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224294, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.530767] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.530887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.531291] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.554502] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d5a3d9-1866-4e48-835a-79b69dc4460f, 'name': SearchDatastore_Task, 'duration_secs': 0.012814} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.555595] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e52b8eb-fa41-4c26-a971-2446fc894ef5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.561692] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 800.561692] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d03532-7422-58c1-d950-50a4bd2f75c4" [ 800.561692] env[61947]: _type = "Task" [ 800.561692] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.571283] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d03532-7422-58c1-d950-50a4bd2f75c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.721202] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Received event network-vif-plugged-46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.721202] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.721202] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.721202] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.721202] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] No waiting events found dispatching network-vif-plugged-46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 800.721753] env[61947]: WARNING nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Received unexpected event network-vif-plugged-46083345-81bb-4fe1-ac51-a4da371e4a16 for instance with vm_state building and task_state spawning. [ 800.721753] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Received event network-changed-46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.721753] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Refreshing instance network info cache due to event network-changed-46083345-81bb-4fe1-ac51-a4da371e4a16. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 800.721753] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.802722] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Successfully created port: 91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.842477] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.875667] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 607d29ad-2db6-4146-a0df-192f727e9d31 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 800.875772] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 683d40c9-b738-486a-806a-7e895637ee85 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 800.875910] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d642bdb3-8e98-4667-9af7-71dae739e6b1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 800.876365] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance c8f37b6f-4f0c-43db-a1ea-3c45d3956f66 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 800.876365] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9a54ca4f-e6ec-4413-b162-fca0cd824e00 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876365] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b7d84a56-edb6-4d1e-b7e9-294078be79d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876365] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 2fc54703-5438-4c2a-b1e7-77431ce5177d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876605] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876605] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876691] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ab027a9d-f51d-4cb6-8aaf-ebd507548657 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.876813] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance cc1f84d4-0a28-48af-88ac-7d977db1cf34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 800.992430] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.001776] env[61947]: DEBUG oslo_vmware.api [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224294, 'name': PowerOnVM_Task, 'duration_secs': 0.540285} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.002159] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 801.002428] env[61947]: INFO nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Took 10.57 seconds to spawn the instance on the hypervisor. [ 801.002660] env[61947]: DEBUG nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.003483] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d5bb79-8255-4288-a057-67e4c5a095d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.078230] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d03532-7422-58c1-d950-50a4bd2f75c4, 'name': SearchDatastore_Task, 'duration_secs': 0.010174} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.078596] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.078954] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb/9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 801.079898] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbee2e5a-98a1-4b09-b346-899a0c91f5a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.087249] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 801.087249] env[61947]: value = "task-1224295" [ 801.087249] env[61947]: _type = "Task" [ 801.087249] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.098662] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224295, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.102147] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.362222] env[61947]: DEBUG nova.network.neutron [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.379699] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.400727] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.401156] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.401575] env[61947]: INFO nova.compute.manager [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Rebooting instance [ 801.523407] env[61947]: INFO nova.compute.manager [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Took 34.19 seconds to build instance. [ 801.599373] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224295, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.857107] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.863968] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.864799] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Instance network_info: |[{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.864799] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.865054] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Refreshing network info cache for port 46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.870173] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:44:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46083345-81bb-4fe1-ac51-a4da371e4a16', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.877992] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating folder: Project (5273b44dfca848538176692ab2a19e3d). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.880549] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d763b9ae-5dec-4005-bddc-55b4832eb015 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.884195] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7efd65d9-67f9-497d-9e42-51d8973aa695 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.890962] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.890962] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.891206] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.891366] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.891513] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.891670] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.891882] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.893567] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.893819] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.893964] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.894169] env[61947]: DEBUG nova.virt.hardware [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.895257] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0efa929-77ca-4e69-8c14-0190c5fc803c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.901868] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created folder: Project (5273b44dfca848538176692ab2a19e3d) in parent group-v264556. [ 801.902034] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating folder: Instances. Parent ref: group-v264604. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.902615] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aca408f-d473-4270-94a9-28a3e528c79e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.913752] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7412dce-7c47-4678-a4da-db84735a61ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.921707] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created folder: Instances in parent group-v264604. [ 801.921992] env[61947]: DEBUG oslo.service.loopingcall [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.930890] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.931957] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d62e6c4-a021-47f3-bc14-62eefcdf7baa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.951610] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.951803] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquired lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.951980] env[61947]: DEBUG nova.network.neutron [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.959217] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.959217] env[61947]: value = "task-1224298" [ 801.959217] env[61947]: _type = "Task" [ 801.959217] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.968299] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224298, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.025603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-49d9579b-36fb-4273-ab73-895f24703f65 tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.102s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.099115] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224295, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534546} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.099464] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb/9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 802.100052] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.101959] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f09e22df-796c-45cc-ac02-a95ab92aa43a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.107147] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 802.107147] env[61947]: value = "task-1224299" [ 802.107147] env[61947]: _type = "Task" [ 802.107147] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.117527] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224299, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.386992] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.469539] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224298, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.529506] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.618064] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224299, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080626} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.618387] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.619233] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39effc14-1dba-4192-aa52-8f27a4b1a929 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.642266] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb/9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.642978] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b34dda73-1f78-423a-888c-759f13146eae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.664234] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 802.664234] env[61947]: value = "task-1224300" [ 802.664234] env[61947]: _type = "Task" [ 802.664234] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.672391] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224300, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.695805] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updated VIF entry in instance network info cache for port 46083345-81bb-4fe1-ac51-a4da371e4a16. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.696174] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.756125] env[61947]: DEBUG nova.compute.manager [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Received event network-vif-plugged-91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 802.756405] env[61947]: DEBUG oslo_concurrency.lockutils [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] Acquiring lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.756619] env[61947]: DEBUG oslo_concurrency.lockutils [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.756791] env[61947]: DEBUG oslo_concurrency.lockutils [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.756991] env[61947]: DEBUG nova.compute.manager [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] No waiting events found dispatching network-vif-plugged-91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.757135] env[61947]: WARNING nova.compute.manager [req-1c4c9186-33f4-40c9-9ac4-a03039c2632a req-7915cad7-a2a5-4220-9cca-ddb912427ce2 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Received unexpected event network-vif-plugged-91399068-64f5-49ce-8e2b-aea7f0be5e47 for instance with vm_state building and task_state spawning. [ 802.890199] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 74e03575-297e-4e08-9236-98d8be80b546 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.971643] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224298, 'name': CreateVM_Task, 'duration_secs': 0.683366} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.974230] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.978166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.978166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.978166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.978166] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41e91f49-8224-4933-b562-332d02d09320 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.981280] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 802.981280] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e33ac2-f8c7-d59c-991a-b166743cd9be" [ 802.981280] env[61947]: _type = "Task" [ 802.981280] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.990303] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e33ac2-f8c7-d59c-991a-b166743cd9be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.036733] env[61947]: DEBUG nova.network.neutron [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [{"id": "6cde2663-2cd8-4148-bb88-3c01384dea78", "address": "fa:16:3e:73:9d:a0", "network": {"id": "9f7fc41f-6e00-4f25-b860-8c1443ca3c9d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-875635745-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634b11b08e6d4c2eb88d7eabe6c8e729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cde2663-2c", "ovs_interfaceid": "6cde2663-2cd8-4148-bb88-3c01384dea78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.044709] env[61947]: DEBUG nova.compute.manager [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Received event network-changed-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.044709] env[61947]: DEBUG nova.compute.manager [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Refreshing instance network info cache due to event network-changed-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 803.044709] env[61947]: DEBUG oslo_concurrency.lockutils [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] Acquiring lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.044990] env[61947]: DEBUG oslo_concurrency.lockutils [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] Acquired lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.044990] env[61947]: DEBUG nova.network.neutron [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Refreshing network info cache for port 2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.062040] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.174310] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224300, 'name': ReconfigVM_Task, 'duration_secs': 0.507626} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.174737] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb/9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.175641] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2245e9f1-417a-43b8-80a7-7f370a0f44f8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.181632] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 803.181632] env[61947]: value = "task-1224301" [ 803.181632] env[61947]: _type = "Task" [ 803.181632] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.194475] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224301, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.199219] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.199565] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.199705] env[61947]: DEBUG nova.compute.manager [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing instance network info cache due to event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 803.199867] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Acquiring lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.393738] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f68a8537-5bae-4b50-b0ff-fa9b4a89bd40 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.459910] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Successfully updated port: 91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.493033] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e33ac2-f8c7-d59c-991a-b166743cd9be, 'name': SearchDatastore_Task, 'duration_secs': 0.022754} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.493033] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.493033] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.493351] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.494044] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.494044] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.494574] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21bbe5a8-2a60-42fb-960f-2f3fe5b8b640 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.503564] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.503843] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 803.504642] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d114ac2-8543-4d10-a68b-66e9ecc1c49c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.510984] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 803.510984] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524298eb-b825-9cbc-dcc0-1c786d95d762" [ 803.510984] env[61947]: _type = "Task" [ 803.510984] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.518570] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524298eb-b825-9cbc-dcc0-1c786d95d762, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.540883] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Releasing lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.542943] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Acquired lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.543074] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing network info cache for port 6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.544573] env[61947]: DEBUG nova.compute.manager [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.545413] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a84f07-cadc-4bca-8f2c-3010e413d0ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.692829] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224301, 'name': Rename_Task, 'duration_secs': 0.169078} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.694046] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.694046] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6422c183-85fc-4207-8a80-faaff5b13ba9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.699982] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 803.699982] env[61947]: value = "task-1224302" [ 803.699982] env[61947]: _type = "Task" [ 803.699982] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.707613] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.897379] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f64b9f7f-89a9-4fb2-9c76-13cc591b9d51 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.962421] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.962585] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquired lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.962729] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.003611] env[61947]: DEBUG nova.network.neutron [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Updated VIF entry in instance network info cache for port 2a7b991a-9fca-4ed2-8096-6317d8bdc7f4. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.003611] env[61947]: DEBUG nova.network.neutron [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Updating instance_info_cache with network_info: [{"id": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "address": "fa:16:3e:81:16:cc", "network": {"id": "cf403b17-d3cc-4a45-91a2-a32571c485eb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1002914311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3a0a834fefd45198f88d7cd28621c5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a7b991a-9f", "ovs_interfaceid": "2a7b991a-9fca-4ed2-8096-6317d8bdc7f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.031703] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524298eb-b825-9cbc-dcc0-1c786d95d762, 'name': SearchDatastore_Task, 'duration_secs': 0.008098} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.033236] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49b08d1a-75c6-4670-84e9-1ac22f8e1179 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.038314] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 804.038314] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f71c12-df36-09fc-3183-d46e4d259c4f" [ 804.038314] env[61947]: _type = "Task" [ 804.038314] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.051056] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f71c12-df36-09fc-3183-d46e4d259c4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.212987] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224302, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.290779] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updated VIF entry in instance network info cache for port 6cde2663-2cd8-4148-bb88-3c01384dea78. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.291336] env[61947]: DEBUG nova.network.neutron [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [{"id": "6cde2663-2cd8-4148-bb88-3c01384dea78", "address": "fa:16:3e:73:9d:a0", "network": {"id": "9f7fc41f-6e00-4f25-b860-8c1443ca3c9d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-875635745-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634b11b08e6d4c2eb88d7eabe6c8e729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cde2663-2c", "ovs_interfaceid": "6cde2663-2cd8-4148-bb88-3c01384dea78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.403530] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.495812] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.506591] env[61947]: DEBUG oslo_concurrency.lockutils [req-b0aef685-8325-4eb0-954f-b050d18d044c req-66bbacdf-2861-4159-bb93-8f87a4b9e8bb service nova] Releasing lock "refresh_cache-7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.549317] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f71c12-df36-09fc-3183-d46e4d259c4f, 'name': SearchDatastore_Task, 'duration_secs': 0.011086} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.549593] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.549851] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 804.550135] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec2950ed-5ae3-4461-9aaa-3b9a5891e4bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.559428] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 804.559428] env[61947]: value = "task-1224303" [ 804.559428] env[61947]: _type = "Task" [ 804.559428] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.564269] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4596beb-ae75-4dac-92eb-181251167bca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.571770] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.573987] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Doing hard reboot of VM {{(pid=61947) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 804.574289] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-35036c93-a4be-44ea-8712-74e560063f24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.581777] env[61947]: DEBUG oslo_vmware.api [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 804.581777] env[61947]: value = "task-1224304" [ 804.581777] env[61947]: _type = "Task" [ 804.581777] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.590647] env[61947]: DEBUG oslo_vmware.api [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224304, 'name': ResetVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.654761] env[61947]: DEBUG nova.network.neutron [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Updating instance_info_cache with network_info: [{"id": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "address": "fa:16:3e:56:98:0f", "network": {"id": "23670e70-5a93-4667-b4fa-c46a762fa17e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1448306464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1edda8c483e448c6a539df24264a14a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91399068-64", "ovs_interfaceid": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.711679] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224302, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.794236] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c9bfb8f-c457-4a1d-a550-b08fb39062f2 req-8bd1c648-ead8-46e5-8571-e8c2006bee21 service nova] Releasing lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.906580] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance eede4f34-5ac4-475c-a74b-a98327f648fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.069054] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224303, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469678} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.069413] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 805.069641] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.070092] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2bfeb73-da4f-4727-810e-eddcd4b6e957 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.075047] env[61947]: DEBUG nova.compute.manager [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Received event network-changed-91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 805.075047] env[61947]: DEBUG nova.compute.manager [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Refreshing instance network info cache due to event network-changed-91399068-64f5-49ce-8e2b-aea7f0be5e47. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 805.075047] env[61947]: DEBUG oslo_concurrency.lockutils [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] Acquiring lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.080162] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 805.080162] env[61947]: value = "task-1224305" [ 805.080162] env[61947]: _type = "Task" [ 805.080162] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.090767] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224305, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.093873] env[61947]: DEBUG oslo_vmware.api [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224304, 'name': ResetVM_Task, 'duration_secs': 0.114145} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.094136] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Did hard reboot of VM {{(pid=61947) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 805.094335] env[61947]: DEBUG nova.compute.manager [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.095715] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c90f6fd-a984-49c7-bd97-28116a8147f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.157552] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Releasing lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.157973] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Instance network_info: |[{"id": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "address": "fa:16:3e:56:98:0f", "network": {"id": "23670e70-5a93-4667-b4fa-c46a762fa17e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1448306464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1edda8c483e448c6a539df24264a14a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91399068-64", "ovs_interfaceid": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.158513] env[61947]: DEBUG oslo_concurrency.lockutils [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] Acquired lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.158749] env[61947]: DEBUG nova.network.neutron [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Refreshing network info cache for port 91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.164020] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:98:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91399068-64f5-49ce-8e2b-aea7f0be5e47', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.173714] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Creating folder: Project (1edda8c483e448c6a539df24264a14a3). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.178097] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12974787-cbe8-4b8b-b745-c542c5cdcd4f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.189779] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Created folder: Project (1edda8c483e448c6a539df24264a14a3) in parent group-v264556. [ 805.191034] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Creating folder: Instances. Parent ref: group-v264607. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.191034] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4b03071-7800-4b0e-bf0f-4c4aa9f941e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.199966] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Created folder: Instances in parent group-v264607. [ 805.200300] env[61947]: DEBUG oslo.service.loopingcall [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.200615] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.200876] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e3e85c7-ba0f-4506-b245-b0222ea3e15c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.229913] env[61947]: DEBUG oslo_vmware.api [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224302, 'name': PowerOnVM_Task, 'duration_secs': 1.472186} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.231207] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.231411] env[61947]: INFO nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Took 9.41 seconds to spawn the instance on the hypervisor. [ 805.231593] env[61947]: DEBUG nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.231895] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.231895] env[61947]: value = "task-1224308" [ 805.231895] env[61947]: _type = "Task" [ 805.231895] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.232630] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e562d7-2868-4734-b34e-f9b9dcf1fe4f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.246491] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224308, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.409809] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7d7f267a-3418-4b9b-9589-c915d5194f0d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.412871] env[61947]: DEBUG nova.network.neutron [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Updated VIF entry in instance network info cache for port 91399068-64f5-49ce-8e2b-aea7f0be5e47. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.413606] env[61947]: DEBUG nova.network.neutron [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Updating instance_info_cache with network_info: [{"id": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "address": "fa:16:3e:56:98:0f", "network": {"id": "23670e70-5a93-4667-b4fa-c46a762fa17e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1448306464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1edda8c483e448c6a539df24264a14a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91399068-64", "ovs_interfaceid": "91399068-64f5-49ce-8e2b-aea7f0be5e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.589488] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224305, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081617} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.589792] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.590588] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2e47b2-9461-4e11-8835-38c82df9761a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.615989] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.616578] env[61947]: DEBUG oslo_concurrency.lockutils [None req-056cc0ea-41fc-4869-95bc-42f328266782 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.216s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.617539] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4acc0cbc-130c-40de-9e64-dff060629315 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.642429] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 805.642429] env[61947]: value = "task-1224309" [ 805.642429] env[61947]: _type = "Task" [ 805.642429] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.650972] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224309, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.744998] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224308, 'name': CreateVM_Task, 'duration_secs': 0.381529} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.745191] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.745865] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.746099] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.746433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.746718] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a543ea63-2c90-441f-be2e-a612af49ba2c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.755270] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 805.755270] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5292c3ac-e2e3-2de9-1c9b-1ce5ad4185cd" [ 805.755270] env[61947]: _type = "Task" [ 805.755270] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.755935] env[61947]: INFO nova.compute.manager [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Took 31.78 seconds to build instance. [ 805.764192] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5292c3ac-e2e3-2de9-1c9b-1ce5ad4185cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.913015] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 72818e40-624b-4c04-888b-bb622f7f96d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.916945] env[61947]: DEBUG oslo_concurrency.lockutils [req-65be8d01-dfa4-4dc7-b09c-8580fa960f64 req-250188c0-b928-40fe-93d3-7d829f17d47f service nova] Releasing lock "refresh_cache-cc1f84d4-0a28-48af-88ac-7d977db1cf34" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.153516] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.262023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c29d1049-e981-412e-9316-cef0c5e698d7 tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.379s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.268373] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5292c3ac-e2e3-2de9-1c9b-1ce5ad4185cd, 'name': SearchDatastore_Task, 'duration_secs': 0.01002} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.268667] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.268983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.269321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.269520] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.269725] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.270012] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df5ef0eb-1285-4cf7-83af-dad3c79f9862 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.279314] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.279558] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.280381] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4c760f8-9d7d-417b-9be8-943fad520812 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.287496] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 806.287496] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52494997-236a-967c-df97-ff1b380f486a" [ 806.287496] env[61947]: _type = "Task" [ 806.287496] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.295137] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52494997-236a-967c-df97-ff1b380f486a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.416522] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.416805] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 806.416954] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 806.656147] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224309, 'name': ReconfigVM_Task, 'duration_secs': 0.781913} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.656506] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfigured VM instance instance-00000034 to attach disk [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.657189] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1adcd552-e6b3-4927-b8c0-fd188e850b8d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.664764] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 806.664764] env[61947]: value = "task-1224310" [ 806.664764] env[61947]: _type = "Task" [ 806.664764] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.672954] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224310, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.737215] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574a1be9-2186-4845-9d3b-a1f0699c5c98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.744593] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d312403d-bdbf-4de0-97e8-38f74b3fd5fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.774051] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.777562] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b8b638-ff58-4359-9b2c-05dac6d392a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.785866] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b73f77-e100-4e53-9fc5-26f49662362f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.802940] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.808477] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52494997-236a-967c-df97-ff1b380f486a, 'name': SearchDatastore_Task, 'duration_secs': 0.009989} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.810170] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba53e521-a657-45af-a775-f313a9454d05 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.816223] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 806.816223] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fa2acf-e545-316a-7516-877e5bfb6dfe" [ 806.816223] env[61947]: _type = "Task" [ 806.816223] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.833201] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fa2acf-e545-316a-7516-877e5bfb6dfe, 'name': SearchDatastore_Task, 'duration_secs': 0.010338} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.833201] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.833372] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] cc1f84d4-0a28-48af-88ac-7d977db1cf34/cc1f84d4-0a28-48af-88ac-7d977db1cf34.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.833620] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36b022bd-81fa-47bf-9849-0d640d1442ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.840119] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 806.840119] env[61947]: value = "task-1224311" [ 806.840119] env[61947]: _type = "Task" [ 806.840119] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.849253] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224311, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.145971] env[61947]: DEBUG nova.compute.manager [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 807.146318] env[61947]: DEBUG nova.compute.manager [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing instance network info cache due to event network-changed-6cde2663-2cd8-4148-bb88-3c01384dea78. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 807.146600] env[61947]: DEBUG oslo_concurrency.lockutils [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] Acquiring lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.146838] env[61947]: DEBUG oslo_concurrency.lockutils [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] Acquired lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.147382] env[61947]: DEBUG nova.network.neutron [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Refreshing network info cache for port 6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.176053] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224310, 'name': Rename_Task, 'duration_secs': 0.188959} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.176439] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 807.176683] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-510d74e5-3f39-4b41-801a-c5589434f05e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.184515] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 807.184515] env[61947]: value = "task-1224312" [ 807.184515] env[61947]: _type = "Task" [ 807.184515] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.193202] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.300592] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.305801] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.328268] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.328643] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.329195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.329458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.329708] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.332324] env[61947]: INFO nova.compute.manager [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Terminating instance [ 807.334247] env[61947]: DEBUG nova.compute.manager [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.334568] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.335275] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6aea6a2-399e-40a7-9d62-649cc699ebe9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.346344] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.346935] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ff969bc-208b-4faa-98c4-e4c7b6616311 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.351447] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224311, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475894} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.352081] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] cc1f84d4-0a28-48af-88ac-7d977db1cf34/cc1f84d4-0a28-48af-88ac-7d977db1cf34.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.352350] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.352639] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b56df6c8-3347-4f08-bc44-1f7f1a9b2bcb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.356083] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 807.356083] env[61947]: value = "task-1224313" [ 807.356083] env[61947]: _type = "Task" [ 807.356083] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.360612] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 807.360612] env[61947]: value = "task-1224314" [ 807.360612] env[61947]: _type = "Task" [ 807.360612] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.369046] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.374858] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224314, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.694991] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224312, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.750356] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.750755] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.750922] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.751137] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.751336] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.757060] env[61947]: INFO nova.compute.manager [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Terminating instance [ 807.759561] env[61947]: DEBUG nova.compute.manager [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.759561] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.761329] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf01e473-85a6-4358-8251-a8b1a0510e49 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.772349] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.772607] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f578c885-76ea-4d51-850b-809faf8b3bbd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.779432] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 807.779432] env[61947]: value = "task-1224315" [ 807.779432] env[61947]: _type = "Task" [ 807.779432] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.791598] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.811442] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 807.811686] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.984s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.813792] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.403s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.815426] env[61947]: INFO nova.compute.claims [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.825308] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 807.826208] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70eca6b9-b703-4745-a020-b5a1bb5ab006 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.834711] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 807.834902] env[61947]: ERROR oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk due to incomplete transfer. [ 807.835149] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4e5626d5-e835-4598-826e-5a1f32db0ba4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.842386] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528503cd-a4fe-af27-ede5-705bddd84117/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 807.842604] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Uploaded image 48c8bce4-a273-4cde-9782-3b9274b22e19 to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 807.845348] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 807.845967] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-01f47704-5e12-4b95-be65-ad353632416e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.854182] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 807.854182] env[61947]: value = "task-1224316" [ 807.854182] env[61947]: _type = "Task" [ 807.854182] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.867900] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224316, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.877862] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224313, 'name': PowerOffVM_Task, 'duration_secs': 0.243071} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.878173] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.878368] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.878869] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d67715fe-488c-47d5-aae9-c85b6a40daa8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.883180] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224314, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070836} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.883720] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.884485] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e49742-e7e3-44ca-8693-6456c4c01794 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.906289] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] cc1f84d4-0a28-48af-88ac-7d977db1cf34/cc1f84d4-0a28-48af-88ac-7d977db1cf34.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.906790] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20f4670f-6ced-4dcc-900d-35904e6c1b2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.922103] env[61947]: DEBUG nova.network.neutron [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updated VIF entry in instance network info cache for port 6cde2663-2cd8-4148-bb88-3c01384dea78. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 807.922463] env[61947]: DEBUG nova.network.neutron [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [{"id": "6cde2663-2cd8-4148-bb88-3c01384dea78", "address": "fa:16:3e:73:9d:a0", "network": {"id": "9f7fc41f-6e00-4f25-b860-8c1443ca3c9d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-875635745-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634b11b08e6d4c2eb88d7eabe6c8e729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cde2663-2c", "ovs_interfaceid": "6cde2663-2cd8-4148-bb88-3c01384dea78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.929261] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 807.929261] env[61947]: value = "task-1224318" [ 807.929261] env[61947]: _type = "Task" [ 807.929261] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.937641] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224318, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.948684] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.948916] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.949150] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Deleting the datastore file [datastore1] 2fc54703-5438-4c2a-b1e7-77431ce5177d {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.949427] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c56b2fdf-7152-454a-94da-de2c3d39a52c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.955964] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for the task: (returnval){ [ 807.955964] env[61947]: value = "task-1224319" [ 807.955964] env[61947]: _type = "Task" [ 807.955964] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.964045] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.195488] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224312, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.290774] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224315, 'name': PowerOffVM_Task, 'duration_secs': 0.342553} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.291053] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 808.291246] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 808.293015] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bea40ea4-283c-434f-b24f-146a1912cb08 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.101984] env[61947]: DEBUG oslo_concurrency.lockutils [req-50d01265-f65d-4d46-97d6-6995894a344a req-6c42afe8-cb59-4563-a6cb-a95c83a35b17 service nova] Releasing lock "refresh_cache-2fc54703-5438-4c2a-b1e7-77431ce5177d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.105145] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224316, 'name': Destroy_Task, 'duration_secs': 0.328282} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.115352] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Destroyed the VM [ 809.116436] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 809.116436] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 809.116436] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 809.116436] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Deleting the datastore file [datastore2] 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.119147] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d34dd31e-df15-41a9-9cc1-c13c33b00baa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.121102] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be4f9174-3f17-4bf9-9262-32be9c3c998c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.132921] env[61947]: DEBUG oslo_vmware.api [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Task: {'id': task-1224319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264561} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.139736] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.139933] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.140123] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.140299] env[61947]: INFO nova.compute.manager [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Took 1.81 seconds to destroy the instance on the hypervisor. [ 809.140556] env[61947]: DEBUG oslo.service.loopingcall [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.141038] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224318, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.141268] env[61947]: DEBUG oslo_vmware.api [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224312, 'name': PowerOnVM_Task, 'duration_secs': 1.687921} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.145043] env[61947]: DEBUG nova.compute.manager [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.145167] env[61947]: DEBUG nova.network.neutron [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.146584] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 809.146784] env[61947]: INFO nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 9.80 seconds to spawn the instance on the hypervisor. [ 809.146994] env[61947]: DEBUG nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.148069] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 809.148069] env[61947]: value = "task-1224321" [ 809.148069] env[61947]: _type = "Task" [ 809.148069] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.150136] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc99cc1-4945-4225-966d-dcc1624d3fc8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.155022] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for the task: (returnval){ [ 809.155022] env[61947]: value = "task-1224322" [ 809.155022] env[61947]: _type = "Task" [ 809.155022] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.169105] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224321, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.172171] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.386577] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 809.386577] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 809.386577] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Cleaning up deleted instances {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 809.426018] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3292884-986e-4b17-a6ac-b85e5e640461 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.433877] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c53c397-5cf6-4e4a-8b7d-380023e4ea5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.470957] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e26b1cf-502d-4d64-b2b6-5c79290e1097 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.480178] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8302773-6318-4786-9d67-b096bf747be1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.499549] env[61947]: DEBUG nova.compute.provider_tree [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.612294] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224318, 'name': ReconfigVM_Task, 'duration_secs': 1.407216} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.612590] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Reconfigured VM instance instance-00000035 to attach disk [datastore2] cc1f84d4-0a28-48af-88ac-7d977db1cf34/cc1f84d4-0a28-48af-88ac-7d977db1cf34.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.613340] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d1ce393-e1cc-4323-b850-027c3a7148de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.618729] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 809.618729] env[61947]: value = "task-1224323" [ 809.618729] env[61947]: _type = "Task" [ 809.618729] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.626951] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.627213] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.627428] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.627619] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.627795] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.629399] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224323, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.630067] env[61947]: INFO nova.compute.manager [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Terminating instance [ 809.632807] env[61947]: DEBUG nova.compute.manager [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.633008] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 809.633784] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3bb728-382a-4197-ae4e-51bedbe94a44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.640831] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 809.641131] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a046621b-0826-4b1a-8191-34a42a7e7f73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.658978] env[61947]: DEBUG nova.compute.manager [req-7f70e0bb-8539-4886-b036-74103af7d6c1 req-19cb14f0-364d-4809-9b45-a0789108bf9e service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Received event network-vif-deleted-6cde2663-2cd8-4148-bb88-3c01384dea78 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.659273] env[61947]: INFO nova.compute.manager [req-7f70e0bb-8539-4886-b036-74103af7d6c1 req-19cb14f0-364d-4809-9b45-a0789108bf9e service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Neutron deleted interface 6cde2663-2cd8-4148-bb88-3c01384dea78; detaching it from the instance and deleting it from the info cache [ 809.659525] env[61947]: DEBUG nova.network.neutron [req-7f70e0bb-8539-4886-b036-74103af7d6c1 req-19cb14f0-364d-4809-9b45-a0789108bf9e service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.674792] env[61947]: DEBUG oslo_vmware.api [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224321, 'name': RemoveSnapshot_Task, 'duration_secs': 0.454758} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.684556] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 809.685327] env[61947]: INFO nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Took 16.11 seconds to snapshot the instance on the hypervisor. [ 809.688711] env[61947]: DEBUG oslo_vmware.api [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Task: {'id': task-1224322, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165508} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.689260] env[61947]: INFO nova.compute.manager [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 25.57 seconds to build instance. [ 809.690411] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.690624] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.690867] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.691090] env[61947]: INFO nova.compute.manager [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Took 1.93 seconds to destroy the instance on the hypervisor. [ 809.691399] env[61947]: DEBUG oslo.service.loopingcall [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.692200] env[61947]: DEBUG nova.compute.manager [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.692345] env[61947]: DEBUG nova.network.neutron [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.701582] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 809.701886] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 809.702122] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleting the datastore file [datastore1] 9a54ca4f-e6ec-4413-b162-fca0cd824e00 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.702433] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40dd7a3b-33de-4c7d-8b85-9d3f8c52b760 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.709171] env[61947]: DEBUG oslo_vmware.api [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 809.709171] env[61947]: value = "task-1224325" [ 809.709171] env[61947]: _type = "Task" [ 809.709171] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.719624] env[61947]: DEBUG oslo_vmware.api [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224325, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.889276] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] There are 4 instances to clean {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 809.889542] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 32fe176d-ebc2-46a4-8e93-3b2555a681d4] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 809.995924] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.996270] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.014145] env[61947]: DEBUG nova.network.neutron [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.032316] env[61947]: ERROR nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [req-fdbfcd59-5ba7-4c80-b064-66fb3961e870] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fdbfcd59-5ba7-4c80-b064-66fb3961e870"}]} [ 810.059406] env[61947]: DEBUG nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 810.079326] env[61947]: DEBUG nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 810.079596] env[61947]: DEBUG nova.compute.provider_tree [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.091059] env[61947]: DEBUG nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 810.110585] env[61947]: DEBUG nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 810.129538] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224323, 'name': Rename_Task, 'duration_secs': 0.190693} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.129812] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.130070] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8fedcdf-a31b-43e0-8e37-0c399ede04cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.137947] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 810.137947] env[61947]: value = "task-1224326" [ 810.137947] env[61947]: _type = "Task" [ 810.137947] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.152960] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.168083] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c24019ac-ed6b-46de-bf8b-8cd6ef250275 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.176296] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96682c21-f1ee-4972-893b-bbed574040b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.191335] env[61947]: DEBUG nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance disappeared during snapshot {{(pid=61947) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 810.194654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c59f560-aff8-4e7e-a2ba-2ce322d4c819 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.292s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.206714] env[61947]: DEBUG nova.compute.manager [req-7f70e0bb-8539-4886-b036-74103af7d6c1 req-19cb14f0-364d-4809-9b45-a0789108bf9e service nova] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Detach interface failed, port_id=6cde2663-2cd8-4148-bb88-3c01384dea78, reason: Instance 2fc54703-5438-4c2a-b1e7-77431ce5177d could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 810.214518] env[61947]: DEBUG nova.compute.manager [None req-1b70a5ac-da46-4ecc-bd75-1495b5814df0 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image not found during clean up 48c8bce4-a273-4cde-9782-3b9274b22e19 {{(pid=61947) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 810.222244] env[61947]: DEBUG oslo_vmware.api [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224325, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213134} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.223832] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.224034] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 810.224277] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 810.224385] env[61947]: INFO nova.compute.manager [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Took 0.59 seconds to destroy the instance on the hypervisor. [ 810.224614] env[61947]: DEBUG oslo.service.loopingcall [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.225207] env[61947]: DEBUG nova.compute.manager [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.225318] env[61947]: DEBUG nova.network.neutron [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.394096] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0c7a6d23-7196-4af2-9f52-6e5a09f26fd8] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 810.443514] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ebf82f-5725-4129-b7aa-9c707984b2e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.451436] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfb99ca-c251-4aaa-8f36-174e83e202ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.497554] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a6a8ef-6a49-4757-b4f4-08ce50ebef32 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.505599] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1a82e0-259a-4f9e-b383-7939fc420f13 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.522129] env[61947]: INFO nova.compute.manager [-] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Took 1.38 seconds to deallocate network for instance. [ 810.523191] env[61947]: DEBUG nova.compute.provider_tree [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.650352] env[61947]: DEBUG oslo_vmware.api [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224326, 'name': PowerOnVM_Task, 'duration_secs': 0.466749} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.650655] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.650867] env[61947]: INFO nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Took 8.79 seconds to spawn the instance on the hypervisor. [ 810.651056] env[61947]: DEBUG nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.651837] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccd2af6-c0ab-44ce-8544-ba5bfecf552c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.672805] env[61947]: DEBUG nova.network.neutron [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.697549] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.897543] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: b35cb8d5-c376-4379-8021-a1e06b09c798] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 811.032627] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.060660] env[61947]: DEBUG nova.scheduler.client.report [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 811.060991] env[61947]: DEBUG nova.compute.provider_tree [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 73 to 74 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 811.061208] env[61947]: DEBUG nova.compute.provider_tree [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.081695] env[61947]: DEBUG nova.network.neutron [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.174810] env[61947]: INFO nova.compute.manager [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Took 25.25 seconds to build instance. [ 811.175807] env[61947]: INFO nova.compute.manager [-] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Took 1.48 seconds to deallocate network for instance. [ 811.220234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.400920] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4a58670f-72ae-48a9-826d-afcff5260047] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 811.567918] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.754s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.568616] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.571688] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.398s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.571897] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.573832] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.070s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.575743] env[61947]: INFO nova.compute.claims [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.586191] env[61947]: INFO nova.compute.manager [-] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Took 1.36 seconds to deallocate network for instance. [ 811.596318] env[61947]: INFO nova.scheduler.client.report [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Deleted allocations for instance d642bdb3-8e98-4667-9af7-71dae739e6b1 [ 811.676748] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6ba2e8d2-027f-47c4-a658-7d53dd741100 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.919s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.685970] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.689094] env[61947]: DEBUG nova.compute.manager [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Received event network-vif-deleted-a6de66aa-4d72-40eb-b4f7-cbdf2d5a3a9d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.691613] env[61947]: DEBUG nova.compute.manager [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Received event network-changed-46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.691613] env[61947]: DEBUG nova.compute.manager [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Refreshing instance network info cache due to event network-changed-46083345-81bb-4fe1-ac51-a4da371e4a16. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.691613] env[61947]: DEBUG oslo_concurrency.lockutils [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.691613] env[61947]: DEBUG oslo_concurrency.lockutils [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.691613] env[61947]: DEBUG nova.network.neutron [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Refreshing network info cache for port 46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.786635] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.786858] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.787086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.787285] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.787458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.789640] env[61947]: INFO nova.compute.manager [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Terminating instance [ 811.791468] env[61947]: DEBUG nova.compute.manager [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 811.791725] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 811.792597] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b39368-44a3-43cc-aed5-97b79459a7c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.801691] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 811.801969] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00430c07-d759-490f-8c75-f8e47fb452d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.808927] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 811.808927] env[61947]: value = "task-1224327" [ 811.808927] env[61947]: _type = "Task" [ 811.808927] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.817491] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.904316] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 811.904605] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Cleaning up deleted instances with incomplete migration {{(pid=61947) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 812.082780] env[61947]: DEBUG nova.compute.utils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.084297] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.084473] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.091632] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.103208] env[61947]: DEBUG oslo_concurrency.lockutils [None req-76c3f9e8-51e3-46c9-9513-00d53a6598c3 tempest-ServerExternalEventsTest-1777568027 tempest-ServerExternalEventsTest-1777568027-project-member] Lock "d642bdb3-8e98-4667-9af7-71dae739e6b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.575s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.133144] env[61947]: DEBUG nova.policy [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65b359bb26c546628af35fddbfcc2d89', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f992ac12ce5244499efac2693a8a517d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 812.181306] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.319756] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224327, 'name': PowerOffVM_Task, 'duration_secs': 0.170207} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.320506] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 812.320735] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 812.321087] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27709482-5534-41ca-bcb8-cb47269f1349 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.407675] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 812.486213] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 812.486465] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 812.486649] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Deleting the datastore file [datastore2] cc1f84d4-0a28-48af-88ac-7d977db1cf34 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.486967] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02b23d9f-d607-4970-add1-984fddfa4d95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.493525] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for the task: (returnval){ [ 812.493525] env[61947]: value = "task-1224329" [ 812.493525] env[61947]: _type = "Task" [ 812.493525] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.501779] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.588602] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.613916] env[61947]: DEBUG nova.network.neutron [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updated VIF entry in instance network info cache for port 46083345-81bb-4fe1-ac51-a4da371e4a16. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.614217] env[61947]: DEBUG nova.network.neutron [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.669850] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Successfully created port: f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.707090] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.958470] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7939d839-598e-4779-bc8a-2fc95e70bfcc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.966976] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e92918e-152c-4595-84b5-bd582405a8f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.001132] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e98ccf-677f-4f4a-89aa-570ca48899c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.008510] env[61947]: DEBUG oslo_vmware.api [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Task: {'id': task-1224329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156609} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.010571] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.010776] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 813.011097] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 813.011332] env[61947]: INFO nova.compute.manager [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Took 1.22 seconds to destroy the instance on the hypervisor. [ 813.011578] env[61947]: DEBUG oslo.service.loopingcall [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.011832] env[61947]: DEBUG nova.compute.manager [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.011928] env[61947]: DEBUG nova.network.neutron [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.014449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa896bab-7993-4e0b-ace4-460b624b4b02 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.028255] env[61947]: DEBUG nova.compute.provider_tree [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.121389] env[61947]: DEBUG oslo_concurrency.lockutils [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.121389] env[61947]: DEBUG nova.compute.manager [req-c976c04b-8bd1-4ec7-b39d-c80721ec7392 req-7b2acbbe-7aa3-4ef8-bf11-0159c529afbf service nova] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Received event network-vif-deleted-c2ea69ed-6dd7-4c61-99aa-c88dd06c8fc2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 813.560941] env[61947]: DEBUG nova.scheduler.client.report [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 74 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 813.561272] env[61947]: DEBUG nova.compute.provider_tree [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 74 to 75 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 813.561596] env[61947]: DEBUG nova.compute.provider_tree [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.600703] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.627077] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.627378] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.627545] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.627730] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.627879] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.628039] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.628290] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.628462] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.628675] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.628863] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.629049] env[61947]: DEBUG nova.virt.hardware [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.629954] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d647a5-14c5-46e5-9afc-ef6ed795e8e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.638341] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6431a289-d61a-4e0f-81af-2c3342faeec8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.716965] env[61947]: DEBUG nova.compute.manager [req-0ab8d262-25d4-4de0-b4be-93bcbed387f0 req-9703ac76-1c0c-4a35-be84-29bfd552ea27 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Received event network-vif-deleted-91399068-64f5-49ce-8e2b-aea7f0be5e47 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 813.717300] env[61947]: INFO nova.compute.manager [req-0ab8d262-25d4-4de0-b4be-93bcbed387f0 req-9703ac76-1c0c-4a35-be84-29bfd552ea27 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Neutron deleted interface 91399068-64f5-49ce-8e2b-aea7f0be5e47; detaching it from the instance and deleting it from the info cache [ 813.717587] env[61947]: DEBUG nova.network.neutron [req-0ab8d262-25d4-4de0-b4be-93bcbed387f0 req-9703ac76-1c0c-4a35-be84-29bfd552ea27 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.887922] env[61947]: DEBUG nova.network.neutron [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.911487] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 813.911653] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 813.911767] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 814.066947] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.067528] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.071831] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.207s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.071946] env[61947]: INFO nova.compute.claims [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.220886] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e909af73-807e-45a6-a626-0fc81ea2c348 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.230767] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ceb3e57-21c4-4cc1-81a0-8ee10b5d3e8e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.258276] env[61947]: DEBUG nova.compute.manager [req-0ab8d262-25d4-4de0-b4be-93bcbed387f0 req-9703ac76-1c0c-4a35-be84-29bfd552ea27 service nova] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Detach interface failed, port_id=91399068-64f5-49ce-8e2b-aea7f0be5e47, reason: Instance cc1f84d4-0a28-48af-88ac-7d977db1cf34 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 814.390895] env[61947]: INFO nova.compute.manager [-] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Took 1.38 seconds to deallocate network for instance. [ 814.417059] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Skipping network cache update for instance because it is being deleted. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 814.417059] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 814.417248] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 814.439971] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.440209] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.440382] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 814.440568] env[61947]: DEBUG nova.objects.instance [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lazy-loading 'info_cache' on Instance uuid 607d29ad-2db6-4146-a0df-192f727e9d31 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.539621] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Successfully updated port: f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.579032] env[61947]: DEBUG nova.compute.utils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.588502] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.588502] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.630504] env[61947]: DEBUG nova.policy [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 814.897956] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.953543] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Successfully created port: 8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.046454] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.046454] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.046454] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.087314] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.442833] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335816da-79ec-4f6e-b8dc-321db364fe5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.459023] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02869e53-03fe-4036-b6ea-82e793903315 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.490318] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.492218] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593cb342-bd01-48f1-a51d-fe9a1850fc25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.501894] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf3b552-d856-4366-b122-9dc464363a61 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.518125] env[61947]: DEBUG nova.compute.provider_tree [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.574394] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.705227] env[61947]: DEBUG nova.network.neutron [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.741513] env[61947]: DEBUG nova.compute.manager [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 815.741721] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.741930] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.742139] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.742285] env[61947]: DEBUG nova.compute.manager [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] No waiting events found dispatching network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.742456] env[61947]: WARNING nova.compute.manager [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received unexpected event network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 for instance with vm_state building and task_state spawning. [ 815.742619] env[61947]: DEBUG nova.compute.manager [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 815.742773] env[61947]: DEBUG nova.compute.manager [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing instance network info cache due to event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 815.742944] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.026280] env[61947]: DEBUG nova.scheduler.client.report [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.074189] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.099974] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.131970] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.131970] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.131970] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.132209] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.132209] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.132209] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.133280] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.133517] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.133724] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.133957] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.135027] env[61947]: DEBUG nova.virt.hardware [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.135207] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de0301f-6fc8-4aa1-b3fc-ee9489cf46a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.145962] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5021c818-3c39-41e0-9bb4-6b1dffb5a15f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.207659] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.208014] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance network_info: |[{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.208404] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.208596] env[61947]: DEBUG nova.network.neutron [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.209983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:73:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4c8c8fd-baca-4e60-97dc-ff0418d63215', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6236fb9-a6fe-4b8d-948a-0518bada02e3', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.217308] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating folder: Project (f992ac12ce5244499efac2693a8a517d). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.224776] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64c856a6-0a05-4eb9-b17e-a9fb8b3dd49b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.236032] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Created folder: Project (f992ac12ce5244499efac2693a8a517d) in parent group-v264556. [ 816.236032] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating folder: Instances. Parent ref: group-v264610. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.236032] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b20cf11-5de4-4965-9656-0f9675a75822 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.244464] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Created folder: Instances in parent group-v264610. [ 816.244709] env[61947]: DEBUG oslo.service.loopingcall [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.244940] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.245135] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f7f6f3a-ef04-4f0c-a810-2fe0b7894f7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.267557] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.267557] env[61947]: value = "task-1224332" [ 816.267557] env[61947]: _type = "Task" [ 816.267557] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.280823] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224332, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.537141] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.537141] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.542046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.761s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.542519] env[61947]: DEBUG nova.objects.instance [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61947) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 816.581196] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-607d29ad-2db6-4146-a0df-192f727e9d31" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.581196] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 816.581196] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.581196] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.581196] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.581196] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.581468] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.581901] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.582291] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 816.582833] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.614155] env[61947]: DEBUG nova.network.neutron [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updated VIF entry in instance network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.614155] env[61947]: DEBUG nova.network.neutron [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.716807] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Successfully updated port: 8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.777743] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224332, 'name': CreateVM_Task, 'duration_secs': 0.297935} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.777915] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.778690] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.778866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.779252] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.779515] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20fdc9ad-9672-4d8d-b651-71921017364d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.784071] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 816.784071] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c30194-f614-86c4-f7dd-d66c2165bd07" [ 816.784071] env[61947]: _type = "Task" [ 816.784071] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.792346] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c30194-f614-86c4-f7dd-d66c2165bd07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.054174] env[61947]: DEBUG nova.compute.utils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.061111] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.061111] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.085750] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.108167] env[61947]: DEBUG nova.policy [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd1ffc62b391469da9bbdfa2763e07e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '025e8ea7833e47e49562ef9ea3bb50a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 817.121876] env[61947]: DEBUG oslo_concurrency.lockutils [req-07383dde-0ec2-4106-bbad-ba7fce5e294b req-09bacded-326b-4775-93c5-5e0d5b98d74c service nova] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.219952] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.220204] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.220757] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.295019] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c30194-f614-86c4-f7dd-d66c2165bd07, 'name': SearchDatastore_Task, 'duration_secs': 0.009042} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.295587] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.295850] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.296087] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.296268] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.296608] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.296939] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-760a0d1f-347f-47d6-9bf1-e2b121e4b62a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.305686] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.305768] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.306483] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0a1aaa3-c96c-47fc-a26f-6d9b53a003de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.311754] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 817.311754] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52880519-7698-b864-6db2-9f8577fbe13e" [ 817.311754] env[61947]: _type = "Task" [ 817.311754] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.319507] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52880519-7698-b864-6db2-9f8577fbe13e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.450798] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully created port: 107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.563685] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.571170] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e62dae8-ac1f-4a3d-af52-cee21cd42007 tempest-ServersAdmin275Test-1698389062 tempest-ServersAdmin275Test-1698389062-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.031s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.571170] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.899s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.575210] env[61947]: INFO nova.compute.claims [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.731865] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully created port: 5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.774685] env[61947]: DEBUG nova.compute.manager [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Received event network-vif-plugged-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.774934] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.775153] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.775331] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.775501] env[61947]: DEBUG nova.compute.manager [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] No waiting events found dispatching network-vif-plugged-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.775673] env[61947]: WARNING nova.compute.manager [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Received unexpected event network-vif-plugged-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe for instance with vm_state building and task_state spawning. [ 817.775974] env[61947]: DEBUG nova.compute.manager [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Received event network-changed-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.775974] env[61947]: DEBUG nova.compute.manager [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Refreshing instance network info cache due to event network-changed-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 817.776286] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Acquiring lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.782135] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.822178] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52880519-7698-b864-6db2-9f8577fbe13e, 'name': SearchDatastore_Task, 'duration_secs': 0.00881} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.822979] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e16a4ad-9b56-4875-88f2-4a58df57893c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.831468] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 817.831468] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52708a6a-bfec-60b9-f5d3-d1505257c079" [ 817.831468] env[61947]: _type = "Task" [ 817.831468] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.843246] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52708a6a-bfec-60b9-f5d3-d1505257c079, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.039193] env[61947]: DEBUG nova.network.neutron [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Updating instance_info_cache with network_info: [{"id": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "address": "fa:16:3e:ae:5c:0f", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8477cd29-b8", "ovs_interfaceid": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.158875] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully created port: b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.342965] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52708a6a-bfec-60b9-f5d3-d1505257c079, 'name': SearchDatastore_Task, 'duration_secs': 0.018436} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.343301] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.343572] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.343841] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7aec2551-4d45-475f-ab42-6f19f13dea1e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.351239] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 818.351239] env[61947]: value = "task-1224333" [ 818.351239] env[61947]: _type = "Task" [ 818.351239] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.359709] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.541942] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.541942] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Instance network_info: |[{"id": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "address": "fa:16:3e:ae:5c:0f", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8477cd29-b8", "ovs_interfaceid": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.542586] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Acquired lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.542724] env[61947]: DEBUG nova.network.neutron [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Refreshing network info cache for port 8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.544248] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:5c:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.553533] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating folder: Project (2367a725fa784cc9bbe918bcdf4b12e3). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.554975] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-109a1543-3f7c-458e-bc9e-3ec54d4243d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.567421] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created folder: Project (2367a725fa784cc9bbe918bcdf4b12e3) in parent group-v264556. [ 818.568894] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating folder: Instances. Parent ref: group-v264613. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.568894] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5af8559-a07d-4d4e-bb70-6a4efdf968cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.579888] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.581029] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created folder: Instances in parent group-v264613. [ 818.581029] env[61947]: DEBUG oslo.service.loopingcall [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.581465] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 818.581668] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88cf53a7-8cc4-4a14-be38-4b6cc34e4102 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.607385] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.607385] env[61947]: value = "task-1224336" [ 818.607385] env[61947]: _type = "Task" [ 818.607385] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.612970] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.613275] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.613473] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.613683] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.613854] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.614049] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.614266] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.614425] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.615697] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.615697] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.615697] env[61947]: DEBUG nova.virt.hardware [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.618035] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26474caa-6aad-411b-8513-882342b03904 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.625744] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224336, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.632195] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb72604-4a39-4b75-bd63-9b1566bc9783 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.863945] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224333, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467066} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.864259] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.864481] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.864741] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9294372-1321-46a0-817a-32808fc67f02 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.871657] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 818.871657] env[61947]: value = "task-1224337" [ 818.871657] env[61947]: _type = "Task" [ 818.871657] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.884401] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.999455] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddafb76-2d4f-4aa1-9b63-6e93ce230c62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.006997] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2052b2b4-521e-4805-ae25-2957b875c03e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.045410] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cded31-2658-4ba0-a3e8-03ab7e5bee12 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.053735] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f98f3c0-95d6-4e7a-ac47-8ecb15479d73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.073018] env[61947]: DEBUG nova.compute.provider_tree [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 819.116138] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224336, 'name': CreateVM_Task, 'duration_secs': 0.354479} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.116332] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 819.117000] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.117181] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.117504] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.117766] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60fdc093-a2b6-4146-aa10-752ab9da7323 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.123999] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 819.123999] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a7e02a-418c-9dd4-2ad8-1f279fa9d323" [ 819.123999] env[61947]: _type = "Task" [ 819.123999] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.132310] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a7e02a-418c-9dd4-2ad8-1f279fa9d323, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.390293] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062812} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.390589] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.391401] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc90fec0-a595-4164-b30c-754f6c0e993f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.414433] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.415341] env[61947]: DEBUG nova.network.neutron [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Updated VIF entry in instance network info cache for port 8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.415682] env[61947]: DEBUG nova.network.neutron [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Updating instance_info_cache with network_info: [{"id": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "address": "fa:16:3e:ae:5c:0f", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8477cd29-b8", "ovs_interfaceid": "8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.416939] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9060bfc6-222d-4cb3-aeb6-fbb1efc25565 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.432324] env[61947]: DEBUG oslo_concurrency.lockutils [req-e8307baa-8068-4ebd-b273-83d6ae17f103 req-18b3be8a-b6c1-4adf-b61e-e2039c0a8d44 service nova] Releasing lock "refresh_cache-7efd65d9-67f9-497d-9e42-51d8973aa695" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.439797] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 819.439797] env[61947]: value = "task-1224338" [ 819.439797] env[61947]: _type = "Task" [ 819.439797] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.449512] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224338, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.635196] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a7e02a-418c-9dd4-2ad8-1f279fa9d323, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.635604] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.635849] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.636309] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.636497] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.636710] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.636985] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a3a3325-230e-4a52-b4b3-26303f1f0145 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.645167] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.645358] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.646118] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-506dca58-e9fd-43db-94df-7fd74ae5c827 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.651183] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 819.651183] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc6129-ddea-4f77-07b8-f0d95b6643bb" [ 819.651183] env[61947]: _type = "Task" [ 819.651183] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.659765] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc6129-ddea-4f77-07b8-f0d95b6643bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.702705] env[61947]: DEBUG nova.scheduler.client.report [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 819.705018] env[61947]: DEBUG nova.compute.provider_tree [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 75 to 76 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 819.705018] env[61947]: DEBUG nova.compute.provider_tree [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 819.877611] env[61947]: DEBUG nova.compute.manager [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-plugged-107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.877844] env[61947]: DEBUG oslo_concurrency.lockutils [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.878014] env[61947]: DEBUG oslo_concurrency.lockutils [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.878993] env[61947]: DEBUG oslo_concurrency.lockutils [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.878993] env[61947]: DEBUG nova.compute.manager [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] No waiting events found dispatching network-vif-plugged-107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.878993] env[61947]: WARNING nova.compute.manager [req-d6bd6421-23c7-4a1d-92ed-08a7cfd7b3e3 req-5a31ee2d-d9ba-424e-9158-f0a5ba82ecf2 service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received unexpected event network-vif-plugged-107026b5-289b-4271-b3e9-053e7aa9e6d6 for instance with vm_state building and task_state spawning. [ 819.881639] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully updated port: 107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.951747] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224338, 'name': ReconfigVM_Task, 'duration_secs': 0.297578} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.953162] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Reconfigured VM instance instance-00000036 to attach disk [datastore2] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.953891] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d32f6de4-fe41-4901-afaf-4ea30d4c9530 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.961744] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 819.961744] env[61947]: value = "task-1224339" [ 819.961744] env[61947]: _type = "Task" [ 819.961744] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.974168] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224339, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.165221] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc6129-ddea-4f77-07b8-f0d95b6643bb, 'name': SearchDatastore_Task, 'duration_secs': 0.01088} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.165995] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5029765-0aa5-4a02-8cf5-da25c8da2b7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.171382] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 820.171382] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5246259b-969c-6337-b264-bb2e043551ac" [ 820.171382] env[61947]: _type = "Task" [ 820.171382] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.179139] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5246259b-969c-6337-b264-bb2e043551ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.213416] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.214236] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.217049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.329s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.217264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.219687] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.023s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.221304] env[61947]: INFO nova.compute.claims [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.247857] env[61947]: INFO nova.scheduler.client.report [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Deleted allocations for instance c8f37b6f-4f0c-43db-a1ea-3c45d3956f66 [ 820.474724] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224339, 'name': Rename_Task, 'duration_secs': 0.15855} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.475393] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.475813] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f592d2e-539d-4517-bb16-8fd0f66bf2f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.484624] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 820.484624] env[61947]: value = "task-1224340" [ 820.484624] env[61947]: _type = "Task" [ 820.484624] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.494852] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.684136] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5246259b-969c-6337-b264-bb2e043551ac, 'name': SearchDatastore_Task, 'duration_secs': 0.009346} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.684472] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.684742] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7efd65d9-67f9-497d-9e42-51d8973aa695/7efd65d9-67f9-497d-9e42-51d8973aa695.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 820.685027] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3d20624-33a6-4f2b-b38d-032b94ae7427 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.693213] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 820.693213] env[61947]: value = "task-1224341" [ 820.693213] env[61947]: _type = "Task" [ 820.693213] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.702500] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.726493] env[61947]: DEBUG nova.compute.utils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.730095] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.730431] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.759257] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10bc6ecb-18cc-4eca-a367-4e09ee5e5827 tempest-ServerMetadataTestJSON-608151257 tempest-ServerMetadataTestJSON-608151257-project-member] Lock "c8f37b6f-4f0c-43db-a1ea-3c45d3956f66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.022s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.780022] env[61947]: DEBUG nova.policy [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f52c8dae4c01408894e45955336869d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '680c8191bf0c45038e60728fdfd54433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 820.995846] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224340, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.056646] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.057516] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.198061] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Successfully created port: 2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.205236] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476978} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.205582] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7efd65d9-67f9-497d-9e42-51d8973aa695/7efd65d9-67f9-497d-9e42-51d8973aa695.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.205834] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.206553] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d3f0d82-5cc7-420a-8ba5-9560422f483a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.214614] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 821.214614] env[61947]: value = "task-1224342" [ 821.214614] env[61947]: _type = "Task" [ 821.214614] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.224322] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.239567] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.494730] env[61947]: DEBUG oslo_vmware.api [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224340, 'name': PowerOnVM_Task, 'duration_secs': 0.991928} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.495252] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.495252] env[61947]: INFO nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Took 7.89 seconds to spawn the instance on the hypervisor. [ 821.495400] env[61947]: DEBUG nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.496584] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb892dfe-f96f-4342-9d73-4c0ae1c53786 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.598017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e320841-a5c8-41e2-99ca-3e69bf7b3d12 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.606775] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf609e22-446c-4cf6-be52-713d068b145a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.642312] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6f6885-96c7-4b4a-9768-9c0ca37f869b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.652918] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d685d061-e1e9-4ad9-b7d3-bc333bb10c74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.668287] env[61947]: DEBUG nova.compute.provider_tree [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.723941] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067019} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.724450] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.725316] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25e8413-00d8-442d-ab74-ef1d1eecb189 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.747459] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 7efd65d9-67f9-497d-9e42-51d8973aa695/7efd65d9-67f9-497d-9e42-51d8973aa695.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.751395] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93d81f68-22dc-45c4-85be-ee960f933caf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.771287] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 821.771287] env[61947]: value = "task-1224343" [ 821.771287] env[61947]: _type = "Task" [ 821.771287] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.780421] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.919302] env[61947]: DEBUG nova.compute.manager [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-changed-107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.919521] env[61947]: DEBUG nova.compute.manager [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing instance network info cache due to event network-changed-107026b5-289b-4271-b3e9-053e7aa9e6d6. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.919763] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Acquiring lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.919956] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Acquired lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.920177] env[61947]: DEBUG nova.network.neutron [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing network info cache for port 107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.021461] env[61947]: INFO nova.compute.manager [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Took 35.64 seconds to build instance. [ 822.173152] env[61947]: DEBUG nova.scheduler.client.report [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.247134] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully updated port: 5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.269083] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.281766] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224343, 'name': ReconfigVM_Task, 'duration_secs': 0.324794} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.282725] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 7efd65d9-67f9-497d-9e42-51d8973aa695/7efd65d9-67f9-497d-9e42-51d8973aa695.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.284196] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afdb7cc4-2f28-400d-936d-7ded832ad54c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.291578] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 822.291578] env[61947]: value = "task-1224344" [ 822.291578] env[61947]: _type = "Task" [ 822.291578] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.297631] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.297904] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.298053] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.298249] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.298403] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.298564] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.298823] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.298955] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.299185] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.299365] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.299543] env[61947]: DEBUG nova.virt.hardware [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.300709] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fb1108-e34f-4875-866c-4cbc20661477 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.308088] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224344, 'name': Rename_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.311135] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d07e7a6-c23e-4126-9120-144a0be01b60 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.483361] env[61947]: DEBUG nova.network.neutron [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.524778] env[61947]: DEBUG oslo_concurrency.lockutils [None req-01c9f072-7438-477a-acc2-a549fca289fa tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.762s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.576894] env[61947]: DEBUG nova.network.neutron [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.676951] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.677519] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.680118] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.811s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.680324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.682592] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.697s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.682775] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.684398] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.692s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.685816] env[61947]: INFO nova.compute.claims [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.719574] env[61947]: INFO nova.scheduler.client.report [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Deleted allocations for instance 607d29ad-2db6-4146-a0df-192f727e9d31 [ 822.728421] env[61947]: INFO nova.scheduler.client.report [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Deleted allocations for instance 683d40c9-b738-486a-806a-7e895637ee85 [ 822.802694] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224344, 'name': Rename_Task, 'duration_secs': 0.145053} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.803213] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.803279] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a82ed25f-ed92-4d34-b1bf-3c0089606e2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.809888] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 822.809888] env[61947]: value = "task-1224345" [ 822.809888] env[61947]: _type = "Task" [ 822.809888] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.818777] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.030999] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.079593] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Releasing lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.079866] env[61947]: DEBUG nova.compute.manager [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-plugged-5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.080084] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.080701] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.080701] env[61947]: DEBUG oslo_concurrency.lockutils [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.080701] env[61947]: DEBUG nova.compute.manager [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] No waiting events found dispatching network-vif-plugged-5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.080864] env[61947]: WARNING nova.compute.manager [req-0bdd4e99-4bde-406e-98fe-723892b9f018 req-f7da8a13-c1d9-45f4-819d-32924db60b1d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received unexpected event network-vif-plugged-5e965567-fa73-4714-9afd-0f7fcf64e2f7 for instance with vm_state building and task_state spawning. [ 823.190959] env[61947]: DEBUG nova.compute.utils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.194647] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.194835] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.208336] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Successfully updated port: 2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.233939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c30a0966-4883-4aff-bd27-60e2d7d92a8a tempest-ServersAdmin275Test-2118102876 tempest-ServersAdmin275Test-2118102876-project-member] Lock "607d29ad-2db6-4146-a0df-192f727e9d31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.130s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.241559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea50915b-8ca4-4e71-9867-4f70611ac425 tempest-ListImageFiltersTestJSON-523309055 tempest-ListImageFiltersTestJSON-523309055-project-member] Lock "683d40c9-b738-486a-806a-7e895637ee85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.204s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.291111] env[61947]: DEBUG nova.policy [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05b9a0659cf14e4883a3400d6da44fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d53677e3f024ef38a7868b1d2a82bc4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 823.322055] env[61947]: DEBUG oslo_vmware.api [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224345, 'name': PowerOnVM_Task, 'duration_secs': 0.458138} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.322343] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.322547] env[61947]: INFO nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Took 7.22 seconds to spawn the instance on the hypervisor. [ 823.322731] env[61947]: DEBUG nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.323554] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cd6612-7c06-4cd4-8488-f5821f9ab264 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.563942] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.695587] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.711602] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.711757] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquired lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.711910] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.747409] env[61947]: DEBUG nova.compute.manager [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.747409] env[61947]: DEBUG nova.compute.manager [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing instance network info cache due to event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 823.747409] env[61947]: DEBUG oslo_concurrency.lockutils [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.747409] env[61947]: DEBUG oslo_concurrency.lockutils [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.747409] env[61947]: DEBUG nova.network.neutron [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.844372] env[61947]: INFO nova.compute.manager [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Took 33.37 seconds to build instance. [ 823.872394] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Successfully created port: d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.956061] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-changed-5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.956240] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing instance network info cache due to event network-changed-5e965567-fa73-4714-9afd-0f7fcf64e2f7. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 823.956472] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Acquiring lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.956630] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Acquired lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.956845] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing network info cache for port 5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.040315] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c164c124-a81e-4c3d-b8b9-a9264896f080 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.049015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fdc936-9f1a-4580-9e94-eaccdc20124c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.082718] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f3d6dc-ccd7-4a11-b923-93dfc46cda14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.091087] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d726c8f-bb01-425e-8613-89157736ca4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.109343] env[61947]: DEBUG nova.compute.provider_tree [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.273735] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.348620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe918e32-da49-48ac-83b0-9a1060fb3b1f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.465s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.542365] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.612795] env[61947]: DEBUG nova.scheduler.client.report [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.714335] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.752705] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.754633] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.754633] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.754633] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.754740] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.755319] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.755319] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.755537] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.755741] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.755916] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.756107] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.757200] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdeb734-514b-43ed-bcf5-64e67d259303 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.760901] env[61947]: DEBUG nova.network.neutron [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Updating instance_info_cache with network_info: [{"id": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "address": "fa:16:3e:ce:2b:e4", "network": {"id": "355e2901-3c9a-4b97-bbcf-2433aeea886c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-216845366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "680c8191bf0c45038e60728fdfd54433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5e3ecd-6d", "ovs_interfaceid": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.768953] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e05d0e-f617-46b7-a9f9-9c413bd6377c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.821346] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.851224] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.938576] env[61947]: DEBUG nova.network.neutron [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updated VIF entry in instance network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.940787] env[61947]: DEBUG nova.network.neutron [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.120888] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.126913] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.127458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.066s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.129122] env[61947]: INFO nova.compute.claims [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.226031] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Successfully updated port: b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.266694] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Releasing lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.266998] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Instance network_info: |[{"id": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "address": "fa:16:3e:ce:2b:e4", "network": {"id": "355e2901-3c9a-4b97-bbcf-2433aeea886c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-216845366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "680c8191bf0c45038e60728fdfd54433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5e3ecd-6d", "ovs_interfaceid": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.267469] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:2b:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89f807d9-140f-4a6f-8bce-96795f9482ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d5e3ecd-6d92-4743-adc8-33b57ad6e561', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.278164] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Creating folder: Project (680c8191bf0c45038e60728fdfd54433). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 825.278542] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a64043aa-0fcd-4045-8b1d-f11d744ffa37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.290800] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Created folder: Project (680c8191bf0c45038e60728fdfd54433) in parent group-v264556. [ 825.291020] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Creating folder: Instances. Parent ref: group-v264616. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 825.293060] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb0428a7-b191-49ed-9a5a-46b7c14d0691 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.303223] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Created folder: Instances in parent group-v264616. [ 825.303223] env[61947]: DEBUG oslo.service.loopingcall [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.303223] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 825.303223] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b4c2fb5-6cf4-46b7-8e5e-012863fcf829 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.327749] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Releasing lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.327749] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Received event network-vif-plugged-2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.327749] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Acquiring lock "74e03575-297e-4e08-9236-98d8be80b546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.327749] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Lock "74e03575-297e-4e08-9236-98d8be80b546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.327749] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Lock "74e03575-297e-4e08-9236-98d8be80b546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.327943] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] No waiting events found dispatching network-vif-plugged-2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.327943] env[61947]: WARNING nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Received unexpected event network-vif-plugged-2d5e3ecd-6d92-4743-adc8-33b57ad6e561 for instance with vm_state building and task_state spawning. [ 825.327943] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Received event network-changed-2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.327943] env[61947]: DEBUG nova.compute.manager [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Refreshing instance network info cache due to event network-changed-2d5e3ecd-6d92-4743-adc8-33b57ad6e561. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 825.327943] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Acquiring lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.328361] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Acquired lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.328361] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Refreshing network info cache for port 2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.329376] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.329376] env[61947]: value = "task-1224348" [ 825.329376] env[61947]: _type = "Task" [ 825.329376] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.337550] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224348, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.392835] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.442995] env[61947]: DEBUG oslo_concurrency.lockutils [req-ab8d6493-71fd-47bd-ada4-d7b4cb522f3e req-9578c5e7-0ec8-40b1-9d41-c08fbd229b06 service nova] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.636118] env[61947]: DEBUG nova.compute.utils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.644028] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.644028] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.733224] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.733896] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.734316] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.743281] env[61947]: DEBUG nova.policy [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05b9a0659cf14e4883a3400d6da44fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d53677e3f024ef38a7868b1d2a82bc4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 825.844533] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224348, 'name': CreateVM_Task, 'duration_secs': 0.503448} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.844811] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 825.846013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.846467] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.846877] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.847238] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cb485af-35e5-4b01-a5fb-c95c8504be6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.851898] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 825.851898] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521180e5-03ce-2345-3129-c51691a6715e" [ 825.851898] env[61947]: _type = "Task" [ 825.851898] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.860834] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521180e5-03ce-2345-3129-c51691a6715e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.986214] env[61947]: DEBUG nova.compute.manager [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-plugged-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.986214] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.986214] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.987199] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.987199] env[61947]: DEBUG nova.compute.manager [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] No waiting events found dispatching network-vif-plugged-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.987199] env[61947]: WARNING nova.compute.manager [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received unexpected event network-vif-plugged-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 for instance with vm_state building and task_state spawning. [ 825.987656] env[61947]: DEBUG nova.compute.manager [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-changed-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.988051] env[61947]: DEBUG nova.compute.manager [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing instance network info cache due to event network-changed-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 825.988391] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Acquiring lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.140637] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.349684] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.372593] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521180e5-03ce-2345-3129-c51691a6715e, 'name': SearchDatastore_Task, 'duration_secs': 0.012046} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.375521] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.375832] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.376033] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.376208] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.376408] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.377474] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fca1c847-fcca-4ee9-9244-47bcd9a2469e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.389140] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.389140] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.389140] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7846b19-b867-49e4-a26a-27acc3946089 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.395997] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 826.395997] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ba1df-8307-9459-cb66-2802b55e95ac" [ 826.395997] env[61947]: _type = "Task" [ 826.395997] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.407151] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ba1df-8307-9459-cb66-2802b55e95ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.418094] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Updated VIF entry in instance network info cache for port 2d5e3ecd-6d92-4743-adc8-33b57ad6e561. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.418094] env[61947]: DEBUG nova.network.neutron [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Updating instance_info_cache with network_info: [{"id": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "address": "fa:16:3e:ce:2b:e4", "network": {"id": "355e2901-3c9a-4b97-bbcf-2433aeea886c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-216845366-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "680c8191bf0c45038e60728fdfd54433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89f807d9-140f-4a6f-8bce-96795f9482ee", "external-id": "nsx-vlan-transportzone-762", "segmentation_id": 762, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5e3ecd-6d", "ovs_interfaceid": "2d5e3ecd-6d92-4743-adc8-33b57ad6e561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.447094] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Successfully updated port: d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.522288] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15747e3-168c-4d8d-ae46-6ad5d380ee1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.534369] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a0ae8c-2652-4ae5-8328-b8a0d3539fdf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.571845] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Successfully created port: 1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.574575] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dbee401-497c-455b-bac3-a6c13fcaba49 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.579550] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.579799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.584801] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9925a6-7081-4a9e-9267-a7ae358d254a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.601538] env[61947]: DEBUG nova.compute.provider_tree [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.907841] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ba1df-8307-9459-cb66-2802b55e95ac, 'name': SearchDatastore_Task, 'duration_secs': 0.0409} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.908887] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6325f9a-450f-4ca2-8d2b-483461d6fe66 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.919039] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 826.919039] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528e5e9b-ebf5-b14c-b81e-fa0f2db48061" [ 826.919039] env[61947]: _type = "Task" [ 826.919039] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.925136] env[61947]: DEBUG oslo_concurrency.lockutils [req-5ee2dc2a-4227-4286-8b15-d8fb6fd61b54 req-cb73f5dc-b5d2-48e2-bf70-e5e45a0f768d service nova] Releasing lock "refresh_cache-74e03575-297e-4e08-9236-98d8be80b546" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.931837] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528e5e9b-ebf5-b14c-b81e-fa0f2db48061, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.950667] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.952766] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.952766] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.086158] env[61947]: DEBUG nova.compute.utils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.104566] env[61947]: DEBUG nova.scheduler.client.report [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.161594] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.191646] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.192053] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.192053] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.193540] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.193540] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.193540] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.193540] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.193540] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.194153] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.194153] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.194153] env[61947]: DEBUG nova.virt.hardware [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.194599] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579cbd28-1be6-48a8-b7ba-4abeecdedcaf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.206339] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49092432-8e3f-446a-8ce3-be233e01e031 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.432220] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528e5e9b-ebf5-b14c-b81e-fa0f2db48061, 'name': SearchDatastore_Task, 'duration_secs': 0.043798} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.432220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.432220] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 74e03575-297e-4e08-9236-98d8be80b546/74e03575-297e-4e08-9236-98d8be80b546.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.432220] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd8ced7c-9858-47d6-8eae-33881832b4c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.441018] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 827.441018] env[61947]: value = "task-1224349" [ 827.441018] env[61947]: _type = "Task" [ 827.441018] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.447277] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.513130] env[61947]: DEBUG nova.network.neutron [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updating instance_info_cache with network_info: [{"id": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "address": "fa:16:3e:15:aa:4d", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.62", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap107026b5-28", "ovs_interfaceid": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "address": "fa:16:3e:c6:7d:f0", "network": {"id": "b89479a1-ec15-4ebb-9913-1666ac514951", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1856841649", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e965567-fa", "ovs_interfaceid": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "address": "fa:16:3e:67:8f:3c", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.201", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3ffb700-3e", "ovs_interfaceid": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.515017] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.589040] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.609316] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.609893] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.616788] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.317s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.619842] env[61947]: INFO nova.compute.claims [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.747546] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Updating instance_info_cache with network_info: [{"id": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "address": "fa:16:3e:d6:81:46", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1be0e35-db", "ovs_interfaceid": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.954806] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224349, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.019100] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.021807] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance network_info: |[{"id": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "address": "fa:16:3e:15:aa:4d", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.62", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap107026b5-28", "ovs_interfaceid": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "address": "fa:16:3e:c6:7d:f0", "network": {"id": "b89479a1-ec15-4ebb-9913-1666ac514951", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1856841649", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e965567-fa", "ovs_interfaceid": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "address": "fa:16:3e:67:8f:3c", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.201", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3ffb700-3e", "ovs_interfaceid": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.021807] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Acquired lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.021807] env[61947]: DEBUG nova.network.neutron [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Refreshing network info cache for port b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.023687] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:aa:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '107026b5-289b-4271-b3e9-053e7aa9e6d6', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:7d:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e965567-fa73-4714-9afd-0f7fcf64e2f7', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:8f:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3ffb700-3e1a-4a00-834d-027bf1a6b6b5', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.039051] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Creating folder: Project (025e8ea7833e47e49562ef9ea3bb50a3). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.040514] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74cebc17-ba41-4178-851b-bf0df4c92ec3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.053574] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Created folder: Project (025e8ea7833e47e49562ef9ea3bb50a3) in parent group-v264556. [ 828.053963] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Creating folder: Instances. Parent ref: group-v264619. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.054072] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e8dac43-8c34-4226-a34e-0e66633d5b63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.069148] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Created folder: Instances in parent group-v264619. [ 828.069148] env[61947]: DEBUG oslo.service.loopingcall [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.069148] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.069148] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-711df992-90a5-41d0-8deb-bb732ecca91d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.103443] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.103443] env[61947]: value = "task-1224352" [ 828.103443] env[61947]: _type = "Task" [ 828.103443] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.112303] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224352, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.125584] env[61947]: DEBUG nova.compute.utils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.127138] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.127353] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 828.249825] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.253031] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Instance network_info: |[{"id": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "address": "fa:16:3e:d6:81:46", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1be0e35-db", "ovs_interfaceid": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.253031] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:81:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1be0e35-dba1-4d37-b9ba-f1bcded1949a', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.259929] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Creating folder: Project (8d53677e3f024ef38a7868b1d2a82bc4). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.260692] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8297c734-33a5-4435-b036-4e7dccdf6a12 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.272265] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Created folder: Project (8d53677e3f024ef38a7868b1d2a82bc4) in parent group-v264556. [ 828.272524] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Creating folder: Instances. Parent ref: group-v264622. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.272834] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36a2cd0c-f21f-4c36-94b5-7178ef5ed9c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.283221] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Created folder: Instances in parent group-v264622. [ 828.283476] env[61947]: DEBUG oslo.service.loopingcall [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.283677] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.283887] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b460a58-3054-486d-ad27-a79c89d8d0cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.303885] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.303885] env[61947]: value = "task-1224355" [ 828.303885] env[61947]: _type = "Task" [ 828.303885] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.312309] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224355, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.447132] env[61947]: DEBUG nova.policy [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 828.456285] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6618} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.456628] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 74e03575-297e-4e08-9236-98d8be80b546/74e03575-297e-4e08-9236-98d8be80b546.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.456881] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.457194] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53d3ff80-e0d5-4929-9bd1-ad16b2ed2d1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.464305] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 828.464305] env[61947]: value = "task-1224356" [ 828.464305] env[61947]: _type = "Task" [ 828.464305] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.479040] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.619366] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224352, 'name': CreateVM_Task, 'duration_secs': 0.432473} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.619582] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 828.621453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.622049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.622365] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.623088] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf0d7312-e4bd-41c3-8b86-61746596dc93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.629529] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 828.629529] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ca79e1-151d-0136-13c9-f746f6e705be" [ 828.629529] env[61947]: _type = "Task" [ 828.629529] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.633882] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.650911] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ca79e1-151d-0136-13c9-f746f6e705be, 'name': SearchDatastore_Task, 'duration_secs': 0.010081} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.653111] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.653189] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.653450] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.653597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.654331] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.654964] env[61947]: DEBUG nova.compute.manager [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Received event network-vif-plugged-d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.656027] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Acquiring lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.656027] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.656027] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.656027] env[61947]: DEBUG nova.compute.manager [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] No waiting events found dispatching network-vif-plugged-d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.656374] env[61947]: WARNING nova.compute.manager [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Received unexpected event network-vif-plugged-d1be0e35-dba1-4d37-b9ba-f1bcded1949a for instance with vm_state building and task_state spawning. [ 828.656736] env[61947]: DEBUG nova.compute.manager [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Received event network-changed-d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.656736] env[61947]: DEBUG nova.compute.manager [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Refreshing instance network info cache due to event network-changed-d1be0e35-dba1-4d37-b9ba-f1bcded1949a. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 828.656915] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Acquiring lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.657209] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Acquired lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.657841] env[61947]: DEBUG nova.network.neutron [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Refreshing network info cache for port d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.659324] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e01d6a89-94e0-42ae-84c7-6d1449be0031 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.668487] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.669857] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 828.669857] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27ae6677-e132-4d42-996c-7b182f139207 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.679790] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.680205] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.680536] env[61947]: INFO nova.compute.manager [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Attaching volume bbff3822-2e8c-44ed-9b43-6b365a574705 to /dev/sdb [ 828.682647] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 828.682647] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521f739f-586b-5933-a2f6-fad59af5d8f3" [ 828.682647] env[61947]: _type = "Task" [ 828.682647] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.702674] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521f739f-586b-5933-a2f6-fad59af5d8f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009516} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.704215] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47b18dc4-609d-4c3c-a50d-ee2d0c6ac84a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.723802] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 828.723802] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e085da-577f-7607-cd1a-9e83203f3e3f" [ 828.723802] env[61947]: _type = "Task" [ 828.723802] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.737056] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e085da-577f-7607-cd1a-9e83203f3e3f, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.737988] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.738322] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd/d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 828.739207] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cc30c5-0681-41e0-bdf2-3a1c32ddc912 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.744280] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13295695-6e31-4b20-a3ee-23bff76eae64 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.751844] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8424ac4-fa98-47c5-9219-e98ec5677dff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.756450] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 828.756450] env[61947]: value = "task-1224357" [ 828.756450] env[61947]: _type = "Task" [ 828.756450] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.767022] env[61947]: DEBUG nova.virt.block_device [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Updating existing volume attachment record: a81112ff-e68a-464b-b289-e6b10ea59ae7 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 828.772329] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.817338] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224355, 'name': CreateVM_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.977921] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073996} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.980102] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.981211] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8044d46-76f9-4dc2-b765-d28cf192dcc1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.008706] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 74e03575-297e-4e08-9236-98d8be80b546/74e03575-297e-4e08-9236-98d8be80b546.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.011836] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-503f1771-f84b-4ff1-976d-433eac989197 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.027713] env[61947]: DEBUG nova.network.neutron [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updated VIF entry in instance network info cache for port b3ffb700-3e1a-4a00-834d-027bf1a6b6b5. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.028202] env[61947]: DEBUG nova.network.neutron [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updating instance_info_cache with network_info: [{"id": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "address": "fa:16:3e:15:aa:4d", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.62", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap107026b5-28", "ovs_interfaceid": "107026b5-289b-4271-b3e9-053e7aa9e6d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "address": "fa:16:3e:c6:7d:f0", "network": {"id": "b89479a1-ec15-4ebb-9913-1666ac514951", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1856841649", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e965567-fa", "ovs_interfaceid": "5e965567-fa73-4714-9afd-0f7fcf64e2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "address": "fa:16:3e:67:8f:3c", "network": {"id": "b785c393-9257-4424-9367-a3fba86ecff3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-10412924", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.201", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3ffb700-3e", "ovs_interfaceid": "b3ffb700-3e1a-4a00-834d-027bf1a6b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.036138] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 829.036138] env[61947]: value = "task-1224360" [ 829.036138] env[61947]: _type = "Task" [ 829.036138] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.048669] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224360, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.067740] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c13ac9-07ee-4b64-a8a3-65a568dbb2c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.079219] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4bf43a-e4e3-40ee-8d1c-a7c01db6bb7c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.120592] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a59d40-bf6f-411c-840f-46ece67bfeb9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.129234] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96c32d9-257e-4cd6-b7e7-9f57d4e7c13e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.148038] env[61947]: DEBUG nova.compute.provider_tree [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.208710] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Successfully updated port: 1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.270252] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224357, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.321593] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224355, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.433853] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Successfully created port: 1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.534090] env[61947]: DEBUG oslo_concurrency.lockutils [req-3794df72-772a-4024-8955-421ac5331677 req-d0412bbd-e114-492d-9d5e-4d396526ad5b service nova] Releasing lock "refresh_cache-d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.554876] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.565623] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.568350] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.625846] env[61947]: DEBUG nova.network.neutron [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Updated VIF entry in instance network info cache for port d1be0e35-dba1-4d37-b9ba-f1bcded1949a. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.626301] env[61947]: DEBUG nova.network.neutron [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Updating instance_info_cache with network_info: [{"id": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "address": "fa:16:3e:d6:81:46", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1be0e35-db", "ovs_interfaceid": "d1be0e35-dba1-4d37-b9ba-f1bcded1949a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.652080] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.660772] env[61947]: DEBUG nova.scheduler.client.report [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.692035] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.692338] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.692505] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.692693] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.692857] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.696456] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.696734] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.696917] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.697116] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.697294] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.698024] env[61947]: DEBUG nova.virt.hardware [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.698756] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e949503e-e84f-4bab-8e7a-3ad898d3bb7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.709616] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f167b6b-3e01-4092-a899-9b9ba6c929ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.714341] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.714567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.714753] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.770307] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705671} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.770307] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd/d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 829.770307] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.770307] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cc9ef3c-b1e5-4a6f-b424-0af7dbe88012 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.776644] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 829.776644] env[61947]: value = "task-1224362" [ 829.776644] env[61947]: _type = "Task" [ 829.776644] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.785463] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224362, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.816300] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224355, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.053717] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224360, 'name': ReconfigVM_Task, 'duration_secs': 0.56692} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.054202] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 74e03575-297e-4e08-9236-98d8be80b546/74e03575-297e-4e08-9236-98d8be80b546.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.055051] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf23e346-d534-4d78-ae5e-a72802b4b5a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.065140] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 830.065140] env[61947]: value = "task-1224363" [ 830.065140] env[61947]: _type = "Task" [ 830.065140] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.076312] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224363, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.135128] env[61947]: DEBUG oslo_concurrency.lockutils [req-e688471f-21c0-44f7-a2e5-beecd963b87c req-5cb3192e-b3bf-44a5-a0c7-777eba1dc9c1 service nova] Releasing lock "refresh_cache-f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.166294] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.166960] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.174964] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.142s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.174964] env[61947]: DEBUG nova.objects.instance [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lazy-loading 'resources' on Instance uuid 2fc54703-5438-4c2a-b1e7-77431ce5177d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.265310] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.292219] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224362, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066686} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.292966] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.293627] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f98368-4e37-4604-a75b-70fc5096361a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.325176] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd/d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.332088] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-234aa9eb-f888-4a8d-8838-66fbb0c04603 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.360617] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224355, 'name': CreateVM_Task, 'duration_secs': 1.73148} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.361919] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.362628] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 830.362628] env[61947]: value = "task-1224364" [ 830.362628] env[61947]: _type = "Task" [ 830.362628] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.363734] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.364056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.364533] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.364981] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b2e88ad-df7e-491b-9dc5-35263524d95a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.373966] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 830.373966] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254edf0-394b-9df5-ac6c-1fe87dfbc605" [ 830.373966] env[61947]: _type = "Task" [ 830.373966] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.378364] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.388318] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254edf0-394b-9df5-ac6c-1fe87dfbc605, 'name': SearchDatastore_Task, 'duration_secs': 0.0103} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.388486] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.388702] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.388943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.389351] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.389351] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.389653] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0558a68f-92da-46fd-848b-d483cec5b2ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.402881] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.402881] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.403168] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cf1e63c-8cd6-4554-96c1-71e77d269446 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.409780] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 830.409780] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f19e1e-436a-09ec-2c43-cc96e70f4f68" [ 830.409780] env[61947]: _type = "Task" [ 830.409780] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.416915] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f19e1e-436a-09ec-2c43-cc96e70f4f68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.514517] env[61947]: DEBUG nova.network.neutron [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Updating instance_info_cache with network_info: [{"id": "1f213945-0710-459d-858f-c5e6ef01a82d", "address": "fa:16:3e:34:e3:97", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f213945-07", "ovs_interfaceid": "1f213945-0710-459d-858f-c5e6ef01a82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.575518] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224363, 'name': Rename_Task, 'duration_secs': 0.148826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.576813] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.576813] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-443bb7c4-f87b-4682-ba94-d68e5c8e157a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.585268] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 830.585268] env[61947]: value = "task-1224365" [ 830.585268] env[61947]: _type = "Task" [ 830.585268] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.593278] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224365, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.676333] env[61947]: DEBUG nova.compute.utils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.677974] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.678344] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.756618] env[61947]: DEBUG nova.policy [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0337baafe604194a5bc93e99ca8078a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51aa57efe97e453783044286f33d3f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 830.877496] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224364, 'name': ReconfigVM_Task, 'duration_secs': 0.272296} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.880640] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Reconfigured VM instance instance-00000038 to attach disk [datastore2] d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd/d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.880640] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2847267f-976c-4d72-8efd-b08fa235f25b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.886059] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 830.886059] env[61947]: value = "task-1224366" [ 830.886059] env[61947]: _type = "Task" [ 830.886059] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.900353] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224366, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.924541] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f19e1e-436a-09ec-2c43-cc96e70f4f68, 'name': SearchDatastore_Task, 'duration_secs': 0.009099} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.925990] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04e164cd-f479-462e-80ee-1c5dc403bd5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.933243] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 830.933243] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52943234-308a-4c26-7fe9-cc01bd5d6d19" [ 830.933243] env[61947]: _type = "Task" [ 830.933243] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.947429] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52943234-308a-4c26-7fe9-cc01bd5d6d19, 'name': SearchDatastore_Task, 'duration_secs': 0.010915} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.947858] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.948531] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] f68a8537-5bae-4b50-b0ff-fa9b4a89bd40/f68a8537-5bae-4b50-b0ff-fa9b4a89bd40.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.948531] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-082281aa-58e2-4288-8868-b822f92b4c35 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.961194] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 830.961194] env[61947]: value = "task-1224367" [ 830.961194] env[61947]: _type = "Task" [ 830.961194] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.969702] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224367, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.017974] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.017974] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Instance network_info: |[{"id": "1f213945-0710-459d-858f-c5e6ef01a82d", "address": "fa:16:3e:34:e3:97", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f213945-07", "ovs_interfaceid": "1f213945-0710-459d-858f-c5e6ef01a82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.018618] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:e3:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f213945-0710-459d-858f-c5e6ef01a82d', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.027246] env[61947]: DEBUG oslo.service.loopingcall [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.030942] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.032531] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69aefb26-0534-413d-8d87-b83a3ed95fee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.058019] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.058019] env[61947]: value = "task-1224368" [ 831.058019] env[61947]: _type = "Task" [ 831.058019] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.064887] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224368, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.086084] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b9c108-e81c-423b-b2e6-85c5d412d274 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.108467] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4498eed-cf3c-4a88-8d95-13a1b1e51f71 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.114362] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224365, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.154037] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf77347-0729-4c5a-89e1-743d98089fac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.164807] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fbe2b9-7d8c-40d7-93e5-a44a29e433b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.170349] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Successfully created port: 5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.183889] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.190035] env[61947]: DEBUG nova.compute.provider_tree [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.263872] env[61947]: DEBUG nova.compute.manager [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Received event network-vif-plugged-1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.264226] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Acquiring lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.266032] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.266032] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.266032] env[61947]: DEBUG nova.compute.manager [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] No waiting events found dispatching network-vif-plugged-1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.266032] env[61947]: WARNING nova.compute.manager [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Received unexpected event network-vif-plugged-1f213945-0710-459d-858f-c5e6ef01a82d for instance with vm_state building and task_state spawning. [ 831.266032] env[61947]: DEBUG nova.compute.manager [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Received event network-changed-1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.266032] env[61947]: DEBUG nova.compute.manager [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Refreshing instance network info cache due to event network-changed-1f213945-0710-459d-858f-c5e6ef01a82d. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 831.266032] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Acquiring lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.266032] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Acquired lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.266032] env[61947]: DEBUG nova.network.neutron [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Refreshing network info cache for port 1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.397689] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224366, 'name': Rename_Task, 'duration_secs': 0.15064} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.398143] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 831.398547] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dae7bb7d-588a-4de7-94ac-89534ad37836 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.406074] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 831.406074] env[61947]: value = "task-1224370" [ 831.406074] env[61947]: _type = "Task" [ 831.406074] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.415787] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.474186] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224367, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.550414] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Successfully updated port: 1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.570603] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224368, 'name': CreateVM_Task, 'duration_secs': 0.357275} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.571535] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.572406] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.572588] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.572909] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.573680] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c293f92b-82d8-40b3-8cf9-bc5cd1200aff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.580052] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 831.580052] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b015-080f-11b1-8d08-135d75c61043" [ 831.580052] env[61947]: _type = "Task" [ 831.580052] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.593205] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b015-080f-11b1-8d08-135d75c61043, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.600086] env[61947]: DEBUG oslo_vmware.api [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224365, 'name': PowerOnVM_Task, 'duration_secs': 0.650127} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.600353] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 831.600775] env[61947]: INFO nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Took 9.33 seconds to spawn the instance on the hypervisor. [ 831.601179] env[61947]: DEBUG nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.602240] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7b0609-6c77-45bf-8407-daf886d18c24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.625171] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "ee3b9509-5437-41b3-b612-91d148a338b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.625344] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.695604] env[61947]: DEBUG nova.scheduler.client.report [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.916837] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224370, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.973700] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224367, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.694246} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.973923] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] f68a8537-5bae-4b50-b0ff-fa9b4a89bd40/f68a8537-5bae-4b50-b0ff-fa9b4a89bd40.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.974159] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.974411] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93d83e0c-d3e3-4c93-813e-03ae3c7a7e59 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.981123] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 831.981123] env[61947]: value = "task-1224371" [ 831.981123] env[61947]: _type = "Task" [ 831.981123] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.988536] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.053102] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.053260] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.053417] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.073162] env[61947]: DEBUG nova.network.neutron [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Updated VIF entry in instance network info cache for port 1f213945-0710-459d-858f-c5e6ef01a82d. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.073162] env[61947]: DEBUG nova.network.neutron [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Updating instance_info_cache with network_info: [{"id": "1f213945-0710-459d-858f-c5e6ef01a82d", "address": "fa:16:3e:34:e3:97", "network": {"id": "c8e108aa-84b4-4b23-b60a-5b12eda5bb28", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1463918001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d53677e3f024ef38a7868b1d2a82bc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f213945-07", "ovs_interfaceid": "1f213945-0710-459d-858f-c5e6ef01a82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.095630] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d7b015-080f-11b1-8d08-135d75c61043, 'name': SearchDatastore_Task, 'duration_secs': 0.054623} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.095995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.096260] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.096506] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.096653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.096836] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.097165] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61fedf59-bfbd-4ec4-8960-8c2cd916c037 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.106178] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.106388] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.107348] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d1d2d8d-180a-41cc-a03e-be4b6c15a127 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.114362] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 832.114362] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526111cc-a716-d8f3-145b-006c588b38c3" [ 832.114362] env[61947]: _type = "Task" [ 832.114362] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.126870] env[61947]: INFO nova.compute.manager [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Took 37.48 seconds to build instance. [ 832.136464] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526111cc-a716-d8f3-145b-006c588b38c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.204021] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.207113] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.032s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.209397] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.989s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.210656] env[61947]: INFO nova.compute.claims [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.234999] env[61947]: INFO nova.scheduler.client.report [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Deleted allocations for instance 2fc54703-5438-4c2a-b1e7-77431ce5177d [ 832.241880] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.242224] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.242488] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.242760] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.243028] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.243291] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.243544] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.243756] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.243975] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.244172] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.244350] env[61947]: DEBUG nova.virt.hardware [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.245636] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443aeed5-13bd-4ef8-9fd6-20f4567ac37d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.254719] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cced0a5c-3353-4804-a34f-9ff0fdabe783 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.419497] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224370, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.491371] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157548} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.491544] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.492431] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca856c5f-02ac-404b-8ece-b63e52239123 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.516425] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] f68a8537-5bae-4b50-b0ff-fa9b4a89bd40/f68a8537-5bae-4b50-b0ff-fa9b4a89bd40.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.516762] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d989fa52-14f5-4a6a-a01a-beb40bc87bd5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.537326] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 832.537326] env[61947]: value = "task-1224372" [ 832.537326] env[61947]: _type = "Task" [ 832.537326] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.545976] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.576069] env[61947]: DEBUG oslo_concurrency.lockutils [req-04e3b230-dcb6-46d5-9af2-aeaab0fb119e req-d9e265cf-c82a-4ab1-b878-6b829a9a8ae5 service nova] Releasing lock "refresh_cache-f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.594311] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.629174] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]526111cc-a716-d8f3-145b-006c588b38c3, 'name': SearchDatastore_Task, 'duration_secs': 0.020604} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.629665] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bbf976a4-d6e0-4272-89c7-28c34a2967f6 tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.354s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.632035] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-132a87ca-f9fd-43e9-b6f9-1721dfe9d865 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.639616] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 832.639616] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c34b4b-5558-0c51-597c-65f92bcec969" [ 832.639616] env[61947]: _type = "Task" [ 832.639616] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.652495] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c34b4b-5558-0c51-597c-65f92bcec969, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.753945] env[61947]: DEBUG oslo_concurrency.lockutils [None req-788eafb5-fb6d-4624-93c4-e234358db0a9 tempest-SecurityGroupsTestJSON-1524362671 tempest-SecurityGroupsTestJSON-1524362671-project-member] Lock "2fc54703-5438-4c2a-b1e7-77431ce5177d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.425s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.825800] env[61947]: DEBUG nova.network.neutron [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.917737] env[61947]: DEBUG oslo_vmware.api [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224370, 'name': PowerOnVM_Task, 'duration_secs': 1.017202} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.917967] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.918257] env[61947]: INFO nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Took 14.34 seconds to spawn the instance on the hypervisor. [ 832.918641] env[61947]: DEBUG nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.919665] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f89ff6c-3749-4a3b-8a6e-ff5be63b3a08 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.048181] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224372, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.135930] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.150095] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c34b4b-5558-0c51-597c-65f92bcec969, 'name': SearchDatastore_Task, 'duration_secs': 0.02859} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.150373] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.150632] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] f64b9f7f-89a9-4fb2-9c76-13cc591b9d51/f64b9f7f-89a9-4fb2-9c76-13cc591b9d51.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.150891] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32bfe23b-0e10-4496-95e0-dd3f2d1c3f6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.158091] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 833.158091] env[61947]: value = "task-1224373" [ 833.158091] env[61947]: _type = "Task" [ 833.158091] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.167960] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224373, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.329127] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.329473] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Instance network_info: |[{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.330158] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:32:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b562a73-9b70-4c31-9d83-f291e5b2cb76', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.338866] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Creating folder: Project (bb7a249773754feaa2ff8a3b9d5bea54). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.339942] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Successfully updated port: 5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.341314] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f936baf3-2875-4e43-80f1-1ae80b4ab900 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.347313] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 833.347635] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264626', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'name': 'volume-bbff3822-2e8c-44ed-9b43-6b365a574705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7efd65d9-67f9-497d-9e42-51d8973aa695', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'serial': 'bbff3822-2e8c-44ed-9b43-6b365a574705'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 833.348551] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380294df-2b32-4731-9637-4d523cc83a83 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.356256] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Created folder: Project (bb7a249773754feaa2ff8a3b9d5bea54) in parent group-v264556. [ 833.356529] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Creating folder: Instances. Parent ref: group-v264628. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.358294] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f702bbc-50ec-4a19-832d-bf246d679a1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.374942] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.375747] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.375747] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.375747] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.375747] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.378140] env[61947]: INFO nova.compute.manager [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Terminating instance [ 833.382034] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311c1905-d2cd-4f2b-b6bf-ba4daef051de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.388567] env[61947]: DEBUG nova.compute.manager [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.388871] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 833.391794] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81c963e-0f25-4200-b839-b6a3f3e96185 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.395384] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Created folder: Instances in parent group-v264628. [ 833.395694] env[61947]: DEBUG oslo.service.loopingcall [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.399216] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.400317] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-301b737b-5271-438c-9922-03c5bffa2d34 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.433144] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.444737] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] volume-bbff3822-2e8c-44ed-9b43-6b365a574705/volume-bbff3822-2e8c-44ed-9b43-6b365a574705.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.447574] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74d30186-f6ed-40df-81df-6b32ff07cb16 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.449857] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b8aa8ac-ad69-4784-ab8a-fad61db0ad6b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.466581] env[61947]: INFO nova.compute.manager [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Took 42.62 seconds to build instance. [ 833.467644] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.467644] env[61947]: value = "task-1224376" [ 833.467644] env[61947]: _type = "Task" [ 833.467644] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.477359] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 833.477359] env[61947]: value = "task-1224377" [ 833.477359] env[61947]: _type = "Task" [ 833.477359] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.477640] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 833.477640] env[61947]: value = "task-1224378" [ 833.477640] env[61947]: _type = "Task" [ 833.477640] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.481118] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224376, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.503101] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224377, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.503775] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224378, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.540235] env[61947]: DEBUG nova.compute.manager [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Received event network-vif-plugged-5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.540486] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.540705] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.540876] env[61947]: DEBUG oslo_concurrency.lockutils [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.541113] env[61947]: DEBUG nova.compute.manager [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] No waiting events found dispatching network-vif-plugged-5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.541249] env[61947]: WARNING nova.compute.manager [req-6c5fdb45-91c0-4c38-97c6-df1f1677682e req-d1b3396d-32f3-4de9-9957-ba52ea22a0da service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Received unexpected event network-vif-plugged-5e37697a-486f-4961-aed8-8ccdbf59992b for instance with vm_state building and task_state spawning. [ 833.555532] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224372, 'name': ReconfigVM_Task, 'duration_secs': 0.564019} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.556844] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Reconfigured VM instance instance-0000003a to attach disk [datastore1] f68a8537-5bae-4b50-b0ff-fa9b4a89bd40/f68a8537-5bae-4b50-b0ff-fa9b4a89bd40.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.557156] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36d5a64b-b0cf-42dc-ac48-47d1aa2e61bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.565353] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 833.565353] env[61947]: value = "task-1224379" [ 833.565353] env[61947]: _type = "Task" [ 833.565353] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.566676] env[61947]: DEBUG nova.compute.manager [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-plugged-1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.566862] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.567108] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.567936] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.567936] env[61947]: DEBUG nova.compute.manager [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] No waiting events found dispatching network-vif-plugged-1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.567936] env[61947]: WARNING nova.compute.manager [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received unexpected event network-vif-plugged-1b562a73-9b70-4c31-9d83-f291e5b2cb76 for instance with vm_state building and task_state spawning. [ 833.567936] env[61947]: DEBUG nova.compute.manager [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-changed-1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.567936] env[61947]: DEBUG nova.compute.manager [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing instance network info cache due to event network-changed-1b562a73-9b70-4c31-9d83-f291e5b2cb76. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 833.568199] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.568267] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.568634] env[61947]: DEBUG nova.network.neutron [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing network info cache for port 1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.586614] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224379, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.656104] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d847a5-f9e1-4d97-b8a6-0299724f15f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.664172] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.668181] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b82a69-d559-4613-9127-dee596e851ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.676207] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224373, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.707543] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa35fe0c-9b7a-434d-b7d3-bffd7eed47b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.716703] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233a5b4e-2b91-49d1-ab14-44dbc0dd39ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.734462] env[61947]: DEBUG nova.compute.provider_tree [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.844361] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.844521] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.844707] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.970370] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8df4a6c4-c269-45ae-8c19-d2346befe755 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.028s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.989032] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224376, 'name': CreateVM_Task, 'duration_secs': 0.434968} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.997849] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 833.999495] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.999673] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.000048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.001208] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31273ca0-5376-4172-8f5b-cc5380e60b55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.009362] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224377, 'name': ReconfigVM_Task, 'duration_secs': 0.520865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.009601] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224378, 'name': PowerOffVM_Task, 'duration_secs': 0.41531} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.010723] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfigured VM instance instance-00000037 to attach disk [datastore1] volume-bbff3822-2e8c-44ed-9b43-6b365a574705/volume-bbff3822-2e8c-44ed-9b43-6b365a574705.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.015557] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.015745] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.016076] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 834.016076] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a3dbd6-87b1-08b6-4fe8-2a4cfa0b7cc0" [ 834.016076] env[61947]: _type = "Task" [ 834.016076] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.016540] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67beb09d-f074-4f1c-925f-c536ecf9b20f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.026428] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a6fab6b-6720-4ba8-b1cd-7423be502ed1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.038511] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a3dbd6-87b1-08b6-4fe8-2a4cfa0b7cc0, 'name': SearchDatastore_Task, 'duration_secs': 0.029158} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.038820] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.039092] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.039353] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.039497] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.039676] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.040993] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c12e75da-b8f6-473b-b294-aaf014c85e44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.049019] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 834.049019] env[61947]: value = "task-1224381" [ 834.049019] env[61947]: _type = "Task" [ 834.049019] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.051434] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.051626] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 834.055490] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72eb20bd-5c6a-43e2-b82e-0cc6cfc97524 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.057880] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.062757] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 834.062757] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522a0f4b-9e68-1f5c-8c3b-d31035385d75" [ 834.062757] env[61947]: _type = "Task" [ 834.062757] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.079470] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522a0f4b-9e68-1f5c-8c3b-d31035385d75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.088809] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224379, 'name': Rename_Task, 'duration_secs': 0.166331} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.089064] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.089367] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51791a01-1993-4c98-a7be-fe5387d7f087 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.097640] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 834.097640] env[61947]: value = "task-1224382" [ 834.097640] env[61947]: _type = "Task" [ 834.097640] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.106884] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.170550] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224373, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531514} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.170834] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] f64b9f7f-89a9-4fb2-9c76-13cc591b9d51/f64b9f7f-89a9-4fb2-9c76-13cc591b9d51.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.171068] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.171339] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8da5b7eb-b366-45dc-9d87-3caa2db79e89 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.178938] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 834.178938] env[61947]: value = "task-1224383" [ 834.178938] env[61947]: _type = "Task" [ 834.178938] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.189199] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.238643] env[61947]: DEBUG nova.scheduler.client.report [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.284346] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.284695] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.284898] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Deleting the datastore file [datastore1] b7d84a56-edb6-4d1e-b7e9-294078be79d8 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.285198] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-452f127b-ec74-473f-8889-6c1c571a3e3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.293895] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for the task: (returnval){ [ 834.293895] env[61947]: value = "task-1224384" [ 834.293895] env[61947]: _type = "Task" [ 834.293895] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.303651] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.397990] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.407092] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.407092] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.407092] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.407465] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.407465] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.410760] env[61947]: INFO nova.compute.manager [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Terminating instance [ 834.415258] env[61947]: DEBUG nova.compute.manager [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.415453] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 834.416359] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69acf58-de48-41c6-b45b-0eb1cd83c267 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.424866] env[61947]: DEBUG nova.network.neutron [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updated VIF entry in instance network info cache for port 1b562a73-9b70-4c31-9d83-f291e5b2cb76. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.425231] env[61947]: DEBUG nova.network.neutron [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.428430] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.428921] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a90df60b-08c0-4b65-acba-3cd49767cc10 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.437180] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 834.437180] env[61947]: value = "task-1224385" [ 834.437180] env[61947]: _type = "Task" [ 834.437180] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.446889] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.477857] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.490630] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "74e03575-297e-4e08-9236-98d8be80b546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.490991] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.491247] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "74e03575-297e-4e08-9236-98d8be80b546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.491462] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.491662] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.500375] env[61947]: INFO nova.compute.manager [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Terminating instance [ 834.502313] env[61947]: DEBUG nova.compute.manager [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.502839] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 834.503468] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135c370e-c187-4417-ace9-968c18a35c00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.521157] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.522760] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5edad4c8-2ea7-47c6-b762-b601fa2cb6ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.534973] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 834.534973] env[61947]: value = "task-1224386" [ 834.534973] env[61947]: _type = "Task" [ 834.534973] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.556012] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224386, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.567790] env[61947]: DEBUG oslo_vmware.api [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224381, 'name': ReconfigVM_Task, 'duration_secs': 0.206649} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.571331] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264626', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'name': 'volume-bbff3822-2e8c-44ed-9b43-6b365a574705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7efd65d9-67f9-497d-9e42-51d8973aa695', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'serial': 'bbff3822-2e8c-44ed-9b43-6b365a574705'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 834.579732] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522a0f4b-9e68-1f5c-8c3b-d31035385d75, 'name': SearchDatastore_Task, 'duration_secs': 0.012426} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.580666] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f11dea52-3374-44dd-ac9c-29b7d1978d15 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.589315] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 834.589315] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52768d4a-c523-d065-d27e-d26569684ccf" [ 834.589315] env[61947]: _type = "Task" [ 834.589315] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.600515] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52768d4a-c523-d065-d27e-d26569684ccf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.609995] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224382, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.626499] env[61947]: DEBUG nova.network.neutron [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.691277] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136563} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.691455] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.692396] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de31b3db-545d-4867-aa96-811c90b40803 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.717147] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] f64b9f7f-89a9-4fb2-9c76-13cc591b9d51/f64b9f7f-89a9-4fb2-9c76-13cc591b9d51.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.717875] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b602f74-21c1-4773-a22c-586455831c24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.740977] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 834.740977] env[61947]: value = "task-1224387" [ 834.740977] env[61947]: _type = "Task" [ 834.740977] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.747724] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.748345] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.753197] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.068s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.753805] env[61947]: DEBUG nova.objects.instance [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lazy-loading 'resources' on Instance uuid 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.765502] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.805629] env[61947]: DEBUG oslo_vmware.api [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Task: {'id': task-1224384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.450876} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.805908] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.806142] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.806314] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.806532] env[61947]: INFO nova.compute.manager [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Took 1.42 seconds to destroy the instance on the hypervisor. [ 834.806880] env[61947]: DEBUG oslo.service.loopingcall [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.811025] env[61947]: DEBUG nova.compute.manager [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.811025] env[61947]: DEBUG nova.network.neutron [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.930313] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd550f70-719c-4df2-90d0-868901a9a34c req-19e34bbe-404b-4dfc-a7e3-6fc6c565bf97 service nova] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.947959] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224385, 'name': PowerOffVM_Task, 'duration_secs': 0.316845} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.948287] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.948480] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.948744] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d14bc206-07e5-48bc-b888-e5a6e9d44b7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.004257] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.046677] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224386, 'name': PowerOffVM_Task, 'duration_secs': 0.35335} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.047008] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.047205] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 835.047463] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a9283fa-e8e1-4895-af57-1c217774966c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.100675] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.100908] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.101122] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleting the datastore file [datastore2] d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.105318] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88e13c2a-8b8b-4f71-a436-a2e194ec7f23 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.112540] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52768d4a-c523-d065-d27e-d26569684ccf, 'name': SearchDatastore_Task, 'duration_secs': 0.034717} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.114165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.114500] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8/04e5cc55-71d9-4d3d-95c1-fb1401ab74f8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 835.115346] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6394b35-e427-488c-a4ad-43a7afdc461d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.123311] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.123620] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.123973] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Deleting the datastore file [datastore1] 74e03575-297e-4e08-9236-98d8be80b546 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.124351] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224382, 'name': PowerOnVM_Task, 'duration_secs': 0.535277} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.126616] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88d49779-f1a4-4000-b8e5-8f80eb378692 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.128568] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.128795] env[61947]: INFO nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Took 10.41 seconds to spawn the instance on the hypervisor. [ 835.128995] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.131240] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 835.131240] env[61947]: value = "task-1224390" [ 835.131240] env[61947]: _type = "Task" [ 835.131240] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.131733] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.132099] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Instance network_info: |[{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.133474] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79aff36-f700-4dc2-a8de-63fb34424348 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.138580] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 835.138580] env[61947]: value = "task-1224391" [ 835.138580] env[61947]: _type = "Task" [ 835.138580] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.141448] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:27:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b931c4c-f73c-4fbd-9c9f-0270834cc69e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e37697a-486f-4961-aed8-8ccdbf59992b', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.155974] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Creating folder: Project (51aa57efe97e453783044286f33d3f00). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.162405] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f1c7a635-fa87-4eb4-a462-975decdf8c4e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.169055] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for the task: (returnval){ [ 835.169055] env[61947]: value = "task-1224392" [ 835.169055] env[61947]: _type = "Task" [ 835.169055] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.185023] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.188584] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.192700] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Created folder: Project (51aa57efe97e453783044286f33d3f00) in parent group-v264556. [ 835.193701] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Creating folder: Instances. Parent ref: group-v264631. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.201023] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6133cf16-711c-405d-bd3f-e22e3a01a166 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.201023] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.213702] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Created folder: Instances in parent group-v264631. [ 835.214183] env[61947]: DEBUG oslo.service.loopingcall [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.217169] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 835.217169] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88e7bf06-02b5-4ae9-b3b8-8140da33ae3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.241942] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.241942] env[61947]: value = "task-1224395" [ 835.241942] env[61947]: _type = "Task" [ 835.241942] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.256448] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224395, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.261662] env[61947]: DEBUG nova.compute.utils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.265305] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224387, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.266102] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 835.654683] env[61947]: DEBUG oslo_vmware.api [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183075} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.656042] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.656280] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 835.656489] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 835.656682] env[61947]: INFO nova.compute.manager [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Took 1.24 seconds to destroy the instance on the hypervisor. [ 835.657084] env[61947]: DEBUG oslo.service.loopingcall [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.657649] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be803d48-8ed6-4f9a-a3cb-c97b7ef47151 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.660334] env[61947]: DEBUG nova.compute.manager [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.660465] env[61947]: DEBUG nova.network.neutron [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.662611] env[61947]: DEBUG nova.objects.instance [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'flavor' on Instance uuid 7efd65d9-67f9-497d-9e42-51d8973aa695 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.679387] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc266d2-70ce-44ca-bd48-0d674ed7788b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.684206] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502725} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.684206] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8/04e5cc55-71d9-4d3d-95c1-fb1401ab74f8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 835.684206] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.684206] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-baa024ea-3656-4f68-b323-b67fcb6d975c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.693440] env[61947]: DEBUG oslo_vmware.api [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Task: {'id': task-1224392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161471} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.730042] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.730042] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 835.730214] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 835.730347] env[61947]: INFO nova.compute.manager [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Took 1.23 seconds to destroy the instance on the hypervisor. [ 835.730604] env[61947]: DEBUG oslo.service.loopingcall [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.730882] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 835.730882] env[61947]: value = "task-1224396" [ 835.730882] env[61947]: _type = "Task" [ 835.730882] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.736018] env[61947]: INFO nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Took 40.57 seconds to build instance. [ 835.736018] env[61947]: DEBUG nova.compute.manager [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.736018] env[61947]: DEBUG nova.network.neutron [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.740125] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e7dfb4-484c-4fc9-b086-05064381f97e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.748982] env[61947]: DEBUG nova.compute.manager [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Received event network-changed-5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.748982] env[61947]: DEBUG nova.compute.manager [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Refreshing instance network info cache due to event network-changed-5e37697a-486f-4961-aed8-8ccdbf59992b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 835.749244] env[61947]: DEBUG oslo_concurrency.lockutils [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] Acquiring lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.749389] env[61947]: DEBUG oslo_concurrency.lockutils [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] Acquired lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.749548] env[61947]: DEBUG nova.network.neutron [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Refreshing network info cache for port 5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.765950] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.769633] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.772432] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224395, 'name': CreateVM_Task, 'duration_secs': 0.424369} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.777130] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.777896] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224387, 'name': ReconfigVM_Task, 'duration_secs': 0.54299} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.778737] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.778912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.779479] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.780469] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fe38ad-6c30-4d0b-b9e2-0e7c74d5d941 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.784268] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Reconfigured VM instance instance-0000003b to attach disk [datastore1] f64b9f7f-89a9-4fb2-9c76-13cc591b9d51/f64b9f7f-89a9-4fb2-9c76-13cc591b9d51.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.784853] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78033094-d64b-47d8-8ae5-4803fcd2a77b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.786576] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b031aae-0b76-437b-a515-a2c8f4fe8529 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.804078] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 835.804078] env[61947]: value = "task-1224397" [ 835.804078] env[61947]: _type = "Task" [ 835.804078] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.806879] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 835.806879] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5247143e-7b20-8430-337a-8a0818a5df47" [ 835.806879] env[61947]: _type = "Task" [ 835.806879] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.806879] env[61947]: DEBUG nova.compute.provider_tree [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.822383] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5247143e-7b20-8430-337a-8a0818a5df47, 'name': SearchDatastore_Task, 'duration_secs': 0.012396} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.824726] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.825011] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.825267] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.825419] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.825685] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.826286] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224397, 'name': Rename_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.827480] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-134c5d33-a7c8-4264-8da2-9676688ed54c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.843211] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.843211] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.843211] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7278bd2e-561d-4e4e-a3ef-ead8c18a11a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.851796] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 835.851796] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f4d02-98f3-dbdb-98bf-915f584d0207" [ 835.851796] env[61947]: _type = "Task" [ 835.851796] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.861121] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f4d02-98f3-dbdb-98bf-915f584d0207, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.170939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a8cddc4c-d805-4e5e-b14b-6cfb4a6d7553 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.491s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.180230] env[61947]: DEBUG nova.compute.manager [req-63bc712f-6b28-401d-b762-688be60b081a req-e18723d2-206a-4971-bd7a-3a137ce90a0c service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Received event network-vif-deleted-ae0ffc38-7ea0-412a-9169-2a2a27dda3eb {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.180478] env[61947]: INFO nova.compute.manager [req-63bc712f-6b28-401d-b762-688be60b081a req-e18723d2-206a-4971-bd7a-3a137ce90a0c service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Neutron deleted interface ae0ffc38-7ea0-412a-9169-2a2a27dda3eb; detaching it from the instance and deleting it from the info cache [ 836.180547] env[61947]: DEBUG nova.network.neutron [req-63bc712f-6b28-401d-b762-688be60b081a req-e18723d2-206a-4971-bd7a-3a137ce90a0c service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.244761] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.156s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.250426] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147245} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.250426] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.250669] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f8ee81-c709-4a43-8e12-6f2b7a74a2bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.276373] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8/04e5cc55-71d9-4d3d-95c1-fb1401ab74f8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.277236] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42b96358-93aa-41b4-9d55-605df242d220 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.302931] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 836.302931] env[61947]: value = "task-1224398" [ 836.302931] env[61947]: _type = "Task" [ 836.302931] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.319237] env[61947]: DEBUG nova.scheduler.client.report [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.330061] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.335154] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.335415] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.335629] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.335807] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.335972] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.337690] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224397, 'name': Rename_Task, 'duration_secs': 0.162392} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.338684] env[61947]: INFO nova.compute.manager [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Terminating instance [ 836.340100] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.341471] env[61947]: DEBUG nova.compute.manager [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.341687] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.341905] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f1b7efe-2923-42c6-b97a-1e0f47073a30 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.343624] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-968de1ad-45cc-4a2c-afff-2ec8fae63dc9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.349945] env[61947]: DEBUG nova.network.neutron [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.355211] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 836.355211] env[61947]: value = "task-1224399" [ 836.355211] env[61947]: _type = "Task" [ 836.355211] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.355811] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 836.355811] env[61947]: value = "task-1224400" [ 836.355811] env[61947]: _type = "Task" [ 836.355811] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.376072] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522f4d02-98f3-dbdb-98bf-915f584d0207, 'name': SearchDatastore_Task, 'duration_secs': 0.010064} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.384412] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224399, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.384412] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.384500] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d2adc6-3f0b-48f6-ae24-2e32eee50e4c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.393111] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 836.393111] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52069e85-a81b-3940-225a-55892237e66c" [ 836.393111] env[61947]: _type = "Task" [ 836.393111] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.405926] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52069e85-a81b-3940-225a-55892237e66c, 'name': SearchDatastore_Task, 'duration_secs': 0.011548} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.406287] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.406487] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.406758] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82f6db55-4ec0-4ea3-a024-f64e463674f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.416364] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 836.416364] env[61947]: value = "task-1224401" [ 836.416364] env[61947]: _type = "Task" [ 836.416364] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.430418] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.565320] env[61947]: DEBUG nova.network.neutron [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.589417] env[61947]: DEBUG nova.network.neutron [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updated VIF entry in instance network info cache for port 5e37697a-486f-4961-aed8-8ccdbf59992b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.589857] env[61947]: DEBUG nova.network.neutron [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.685210] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15d2268a-60a1-42b3-b747-fcfc433a05f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.699077] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3de2e1-c126-4e76-8a81-ad7ea4849765 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.739821] env[61947]: DEBUG nova.compute.manager [req-63bc712f-6b28-401d-b762-688be60b081a req-e18723d2-206a-4971-bd7a-3a137ce90a0c service nova] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Detach interface failed, port_id=ae0ffc38-7ea0-412a-9169-2a2a27dda3eb, reason: Instance b7d84a56-edb6-4d1e-b7e9-294078be79d8 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 836.750850] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 836.796420] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.820114] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224398, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.830141] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.830335] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.830547] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.830782] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.831083] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.831286] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.831456] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.831698] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.831936] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.832142] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.832392] env[61947]: DEBUG nova.virt.hardware [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.833539] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.081s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.836447] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff99dea-f434-449b-a482-1994b9894d65 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.839635] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.748s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.840112] env[61947]: DEBUG nova.objects.instance [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lazy-loading 'resources' on Instance uuid 9a54ca4f-e6ec-4413-b162-fca0cd824e00 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.850139] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b632987-256b-4310-9aed-06b99c2a32df {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.857119] env[61947]: INFO nova.compute.manager [-] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Took 2.05 seconds to deallocate network for instance. [ 836.880379] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.885243] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Creating folder: Project (fe0da8218bd041f6b705744a4f190ae0). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.886635] env[61947]: INFO nova.scheduler.client.report [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Deleted allocations for instance 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb [ 836.893570] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90728431-5fc8-427f-bb76-31bb7dfbb26f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.899647] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224400, 'name': PowerOffVM_Task, 'duration_secs': 0.42234} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.903393] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 836.903463] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 836.903661] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264626', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'name': 'volume-bbff3822-2e8c-44ed-9b43-6b365a574705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7efd65d9-67f9-497d-9e42-51d8973aa695', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'serial': 'bbff3822-2e8c-44ed-9b43-6b365a574705'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 836.904374] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224399, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.905211] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0689ed37-9ffd-4fea-8c12-8ac15fdf049c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.911386] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Created folder: Project (fe0da8218bd041f6b705744a4f190ae0) in parent group-v264556. [ 836.911594] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Creating folder: Instances. Parent ref: group-v264634. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.914065] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b01f96c2-77c7-42ea-85cf-bd6358fbbc73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.938684] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7efd445-9070-49c8-b3ed-72e953a6e3be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.941659] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Created folder: Instances in parent group-v264634. [ 836.941942] env[61947]: DEBUG oslo.service.loopingcall [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.942560] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 836.943753] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-992a62a5-0dd9-4efb-a655-32479102a7da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.964573] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224401, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512829} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.965814] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6131e886-6569-4174-9f51-b32aa97e0a50 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.968379] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 836.968686] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.969264] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e9d3f2e-e745-49f5-bf6a-d0445a2e33d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.973217] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.973217] env[61947]: value = "task-1224404" [ 836.973217] env[61947]: _type = "Task" [ 836.973217] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.999021] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ab0f2a-24a4-4d2e-bdbd-91f1d18d9aea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.000593] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 837.000593] env[61947]: value = "task-1224405" [ 837.000593] env[61947]: _type = "Task" [ 837.000593] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.004965] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224404, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.023500] env[61947]: DEBUG nova.network.neutron [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.026134] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] The volume has not been displaced from its original location: [datastore1] volume-bbff3822-2e8c-44ed-9b43-6b365a574705/volume-bbff3822-2e8c-44ed-9b43-6b365a574705.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 837.036022] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 837.036022] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-755b9e9d-df35-49da-bd89-3a93d451cc7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.053777] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224405, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.060217] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 837.060217] env[61947]: value = "task-1224406" [ 837.060217] env[61947]: _type = "Task" [ 837.060217] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.070055] env[61947]: INFO nova.compute.manager [-] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Took 1.33 seconds to deallocate network for instance. [ 837.070242] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.095051] env[61947]: DEBUG oslo_concurrency.lockutils [req-ceb2fe13-8183-4da1-84d3-fe67a044cc91 req-37f4d7e3-8665-4578-b892-5ca3fab7f32f service nova] Releasing lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.283346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.316604] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224398, 'name': ReconfigVM_Task, 'duration_secs': 0.562618} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.316917] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8/04e5cc55-71d9-4d3d-95c1-fb1401ab74f8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.317662] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3108405e-6f48-46b3-ae17-94a30a9b877d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.326739] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 837.326739] env[61947]: value = "task-1224407" [ 837.326739] env[61947]: _type = "Task" [ 837.326739] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.339578] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224407, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.365414] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.378738] env[61947]: DEBUG oslo_vmware.api [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224399, 'name': PowerOnVM_Task, 'duration_secs': 0.697369} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.379099] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.379801] env[61947]: INFO nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Took 10.22 seconds to spawn the instance on the hypervisor. [ 837.380043] env[61947]: DEBUG nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.380898] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de34e4fd-26e6-46c8-a62a-92cc8610e8ed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.399991] env[61947]: DEBUG oslo_concurrency.lockutils [None req-26252f25-73f6-4c3e-a9ca-72a001b45a2a tempest-ServerTagsTestJSON-568707548 tempest-ServerTagsTestJSON-568707548-project-member] Lock "9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.649s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.485364] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224404, 'name': CreateVM_Task, 'duration_secs': 0.425582} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.485531] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.485963] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.486148] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.486471] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.486726] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccff8082-4c99-4282-88ed-e2f35f5a1384 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.492528] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 837.492528] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd5567-fa3b-a445-d78f-f8dc431f5200" [ 837.492528] env[61947]: _type = "Task" [ 837.492528] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.505165] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd5567-fa3b-a445-d78f-f8dc431f5200, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.517202] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224405, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097617} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.517495] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.518684] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d9a2a7-0717-4bf5-ba45-d3f32a4ff028 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.526371] env[61947]: INFO nova.compute.manager [-] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Took 1.87 seconds to deallocate network for instance. [ 837.549295] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.555580] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0a45a87-b519-412e-b5ee-c55ff24bb41f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.581020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.581831] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224406, 'name': ReconfigVM_Task, 'duration_secs': 0.30096} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.585890] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 837.591639] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 837.591639] env[61947]: value = "task-1224408" [ 837.591639] env[61947]: _type = "Task" [ 837.591639] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.592105] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f14e8d0-6ef9-4d10-ba15-698bb5988990 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.615866] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.617377] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 837.617377] env[61947]: value = "task-1224409" [ 837.617377] env[61947]: _type = "Task" [ 837.617377] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.628929] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.712595] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be2fb12-c9ac-441d-9b67-eb4069c99174 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.721449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0cb6a0-ebd7-47bf-bc7c-02054d7c134b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.756658] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90b3709-b381-4151-97ac-f09d9e66bfa6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.766201] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c6f87e-a0da-4956-b531-fc848e58b005 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.782275] env[61947]: DEBUG nova.compute.provider_tree [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.837052] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224407, 'name': Rename_Task, 'duration_secs': 0.239911} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.837382] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 837.837650] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d22e298-565b-47ad-8624-a098035eea22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.845247] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 837.845247] env[61947]: value = "task-1224410" [ 837.845247] env[61947]: _type = "Task" [ 837.845247] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.857518] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.886641] env[61947]: DEBUG nova.compute.manager [req-9644ad2c-dc2a-4b91-a1cc-297c6b54a45e req-7252bc0b-4522-4ddc-a290-2810c9c72eec service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-deleted-b3ffb700-3e1a-4a00-834d-027bf1a6b6b5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.886948] env[61947]: DEBUG nova.compute.manager [req-9644ad2c-dc2a-4b91-a1cc-297c6b54a45e req-7252bc0b-4522-4ddc-a290-2810c9c72eec service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-deleted-107026b5-289b-4271-b3e9-053e7aa9e6d6 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.887262] env[61947]: DEBUG nova.compute.manager [req-9644ad2c-dc2a-4b91-a1cc-297c6b54a45e req-7252bc0b-4522-4ddc-a290-2810c9c72eec service nova] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Received event network-vif-deleted-5e965567-fa73-4714-9afd-0f7fcf64e2f7 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.911637] env[61947]: INFO nova.compute.manager [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Took 36.94 seconds to build instance. [ 838.012462] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd5567-fa3b-a445-d78f-f8dc431f5200, 'name': SearchDatastore_Task, 'duration_secs': 0.01159} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.013029] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.013417] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.013833] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.014163] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.014492] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.015305] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de00b522-9c52-43c3-95eb-8847b6125397 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.030990] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.032711] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.032711] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e653d3a7-9ff1-4953-8a7d-73e6632ed29a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.044035] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 838.044035] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f7633b-9a5a-8273-ae6d-1ca49474cd43" [ 838.044035] env[61947]: _type = "Task" [ 838.044035] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.055519] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f7633b-9a5a-8273-ae6d-1ca49474cd43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.077153] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.115394] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.127727] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224409, 'name': ReconfigVM_Task, 'duration_secs': 0.382352} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.127876] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264626', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'name': 'volume-bbff3822-2e8c-44ed-9b43-6b365a574705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7efd65d9-67f9-497d-9e42-51d8973aa695', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbff3822-2e8c-44ed-9b43-6b365a574705', 'serial': 'bbff3822-2e8c-44ed-9b43-6b365a574705'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 838.128199] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 838.129027] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a12166-f437-4757-8b86-ced7281c62e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.142412] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 838.142822] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e078269b-ca4f-49f2-97a0-d7bee8104fc3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.288459] env[61947]: DEBUG nova.scheduler.client.report [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.301670] env[61947]: DEBUG nova.compute.manager [req-34433841-8302-443b-b956-63a09c80c64d req-dfb660d1-2675-4ae0-aa37-d45d6c4f5c12 service nova] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Received event network-vif-deleted-2d5e3ecd-6d92-4743-adc8-33b57ad6e561 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.306182] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 838.306566] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 838.307021] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleting the datastore file [datastore2] 7efd65d9-67f9-497d-9e42-51d8973aa695 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 838.307239] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe808e77-4461-4ca6-8d1c-af90a44c2160 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.316210] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 838.316210] env[61947]: value = "task-1224412" [ 838.316210] env[61947]: _type = "Task" [ 838.316210] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.328457] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.355441] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224410, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.414046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6167c0ca-5488-40f6-a1b7-a800cfb3e2a2 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.296s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.553208] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f7633b-9a5a-8273-ae6d-1ca49474cd43, 'name': SearchDatastore_Task, 'duration_secs': 0.018538} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.553208] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d972278-a0ac-4af1-bf49-384fa1e68a6f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.562021] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 838.562021] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e36cf3-cc72-effc-489d-d61e88f39d0e" [ 838.562021] env[61947]: _type = "Task" [ 838.562021] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.568613] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e36cf3-cc72-effc-489d-d61e88f39d0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.614435] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224408, 'name': ReconfigVM_Task, 'duration_secs': 0.593294} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.614775] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfigured VM instance instance-0000003d to attach disk [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.615504] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ffb2929-4996-4a0b-8cb2-598fdad183cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.624219] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 838.624219] env[61947]: value = "task-1224413" [ 838.624219] env[61947]: _type = "Task" [ 838.624219] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.633612] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224413, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.794705] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.955s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.798605] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.090s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.800306] env[61947]: INFO nova.compute.claims [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.830024] env[61947]: DEBUG oslo_vmware.api [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.383832} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.830024] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.830024] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.830024] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.830024] env[61947]: INFO nova.compute.manager [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Took 2.49 seconds to destroy the instance on the hypervisor. [ 838.830024] env[61947]: DEBUG oslo.service.loopingcall [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.830024] env[61947]: DEBUG nova.compute.manager [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.830024] env[61947]: DEBUG nova.network.neutron [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.838217] env[61947]: INFO nova.scheduler.client.report [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted allocations for instance 9a54ca4f-e6ec-4413-b162-fca0cd824e00 [ 838.858333] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224410, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.006706] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.007015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.007247] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.007482] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.007695] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.014007] env[61947]: INFO nova.compute.manager [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Terminating instance [ 839.019807] env[61947]: DEBUG nova.compute.manager [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.019953] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.020888] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f3943a-a956-460e-843c-0fb755b4a3e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.030127] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.030446] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1b2e32a-65ce-461c-91ff-02c61bc9e9f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.040115] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 839.040115] env[61947]: value = "task-1224414" [ 839.040115] env[61947]: _type = "Task" [ 839.040115] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.050948] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.073591] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e36cf3-cc72-effc-489d-d61e88f39d0e, 'name': SearchDatastore_Task, 'duration_secs': 0.03803} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.074204] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.074749] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7d7f267a-3418-4b9b-9589-c915d5194f0d/7d7f267a-3418-4b9b-9589-c915d5194f0d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.075195] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-542dde1b-392b-4036-a16e-4facc6cb93b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.087778] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 839.087778] env[61947]: value = "task-1224415" [ 839.087778] env[61947]: _type = "Task" [ 839.087778] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.098651] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.136289] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224413, 'name': Rename_Task, 'duration_secs': 0.319045} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.136289] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 839.136454] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1498aa5c-3ac1-4ffe-93ee-5182981fe1b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.145657] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 839.145657] env[61947]: value = "task-1224416" [ 839.145657] env[61947]: _type = "Task" [ 839.145657] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.159653] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.351020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad0d01f7-2a9c-44bd-b271-88f818ac7896 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9a54ca4f-e6ec-4413-b162-fca0cd824e00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.723s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.363085] env[61947]: DEBUG oslo_vmware.api [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224410, 'name': PowerOnVM_Task, 'duration_secs': 1.094763} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.363422] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.363662] env[61947]: INFO nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Took 9.71 seconds to spawn the instance on the hypervisor. [ 839.363919] env[61947]: DEBUG nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.364807] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1f4644-79d4-49d3-940e-af3454c57dec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.549211] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224414, 'name': PowerOffVM_Task, 'duration_secs': 0.207538} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.549554] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.549739] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.550037] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0166aa90-5e3f-495b-ba52-c1eda14dc841 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.606512] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224415, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488852} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.607031] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 7d7f267a-3418-4b9b-9589-c915d5194f0d/7d7f267a-3418-4b9b-9589-c915d5194f0d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 839.607031] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.607491] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8956045-eca4-4f5e-aeef-f05e66f2341e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.619201] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 839.619201] env[61947]: value = "task-1224418" [ 839.619201] env[61947]: _type = "Task" [ 839.619201] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.625821] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.626183] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.629931] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Deleting the datastore file [datastore2] 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.629931] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d2efd86-785a-48c3-8e35-059ec5b1616d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.632447] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224418, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.638795] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for the task: (returnval){ [ 839.638795] env[61947]: value = "task-1224419" [ 839.638795] env[61947]: _type = "Task" [ 839.638795] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.650289] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.663618] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224416, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.889214] env[61947]: INFO nova.compute.manager [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Took 36.85 seconds to build instance. [ 840.085410] env[61947]: DEBUG nova.network.neutron [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.135301] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224418, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076992} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.135541] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.136448] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa35eaa-b96d-4b93-bb7f-9ba2f678f92b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.167362] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 7d7f267a-3418-4b9b-9589-c915d5194f0d/7d7f267a-3418-4b9b-9589-c915d5194f0d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.172714] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbfa42ba-36b8-4a2d-9f35-64179d0ef3e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.195650] env[61947]: DEBUG oslo_vmware.api [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Task: {'id': task-1224419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162951} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.197511] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.197618] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.198454] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.198454] env[61947]: INFO nova.compute.manager [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 840.198454] env[61947]: DEBUG oslo.service.loopingcall [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.202210] env[61947]: DEBUG nova.compute.manager [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.202321] env[61947]: DEBUG nova.network.neutron [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.204541] env[61947]: DEBUG oslo_vmware.api [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224416, 'name': PowerOnVM_Task, 'duration_secs': 0.665919} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.205834] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 840.206077] env[61947]: INFO nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Took 8.00 seconds to spawn the instance on the hypervisor. [ 840.206294] env[61947]: DEBUG nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.206683] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 840.206683] env[61947]: value = "task-1224420" [ 840.206683] env[61947]: _type = "Task" [ 840.206683] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.207839] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e426be-2814-4e4b-ab74-856db8df263d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.224833] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.236980] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6eea46-a64c-4cd4-aecc-4208ddede777 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.246805] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a02d52-247f-4978-b24d-03c8c1b9e67f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.283259] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9584e9e-4597-4abc-8966-88729341e82a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.291921] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816d950d-3de2-4dc9-883c-e8c6f41e7a1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.307685] env[61947]: DEBUG nova.compute.provider_tree [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.309155] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.309523] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.309611] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.309818] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.309989] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.312980] env[61947]: INFO nova.compute.manager [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Terminating instance [ 840.317162] env[61947]: DEBUG nova.compute.manager [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.317162] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.317162] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afca91b3-86fb-48a7-943e-e92df0ca580a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.325169] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.325432] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300f4742-8b4e-4854-94cd-d6a7d235fd28 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.334262] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 840.334262] env[61947]: value = "task-1224421" [ 840.334262] env[61947]: _type = "Task" [ 840.334262] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.344570] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224421, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.371953] env[61947]: DEBUG nova.compute.manager [req-76fc5fd3-9704-4719-81a5-4841368e2e09 req-036e60b4-ffb6-4b9f-a86a-e69580d75779 service nova] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Received event network-vif-deleted-8477cd29-b8ea-4e8b-86f0-5bbad7c7f5fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.392208] env[61947]: DEBUG oslo_concurrency.lockutils [None req-84aa488a-69fc-47e9-8900-f4348e936156 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.297s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.401869] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.402158] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.402375] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.402618] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.402797] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.405427] env[61947]: INFO nova.compute.manager [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Terminating instance [ 840.407473] env[61947]: DEBUG nova.compute.manager [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.407653] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.408537] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1912090e-fc38-4939-bd32-bc81f00b121e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.417588] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.417807] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-747464ea-02f8-45ac-8931-c91c6153a1ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.426569] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 840.426569] env[61947]: value = "task-1224422" [ 840.426569] env[61947]: _type = "Task" [ 840.426569] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.438713] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.593567] env[61947]: INFO nova.compute.manager [-] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Took 1.76 seconds to deallocate network for instance. [ 840.721848] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.736354] env[61947]: INFO nova.compute.manager [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Took 33.45 seconds to build instance. [ 840.812909] env[61947]: DEBUG nova.scheduler.client.report [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.850413] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224421, 'name': PowerOffVM_Task, 'duration_secs': 0.245792} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.850413] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.850413] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.850413] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a296ed6-ce6b-496e-948e-4a18666846bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.919767] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.919767] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.919767] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleting the datastore file [datastore1] f68a8537-5bae-4b50-b0ff-fa9b4a89bd40 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.919767] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbffaf9d-d3f6-41fa-a200-91edf7f80644 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.931941] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 840.931941] env[61947]: value = "task-1224424" [ 840.931941] env[61947]: _type = "Task" [ 840.931941] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.942413] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224422, 'name': PowerOffVM_Task, 'duration_secs': 0.423204} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.942413] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.942413] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.942413] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87c0fe65-a99f-496f-9996-8f7f6743f285 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.951474] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.022880] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 841.023183] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 841.023388] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleting the datastore file [datastore1] f64b9f7f-89a9-4fb2-9c76-13cc591b9d51 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.023701] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc8986d4-464b-477b-be59-88242a22aca4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.033262] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for the task: (returnval){ [ 841.033262] env[61947]: value = "task-1224426" [ 841.033262] env[61947]: _type = "Task" [ 841.033262] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.044631] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224426, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.185081] env[61947]: INFO nova.compute.manager [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Took 0.59 seconds to detach 1 volumes for instance. [ 841.227830] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224420, 'name': ReconfigVM_Task, 'duration_secs': 0.693431} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.228171] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 7d7f267a-3418-4b9b-9589-c915d5194f0d/7d7f267a-3418-4b9b-9589-c915d5194f0d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.228824] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b8dd6fb-3a4a-48fb-b392-40cd57e0b94d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.238457] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 841.238457] env[61947]: value = "task-1224427" [ 841.238457] env[61947]: _type = "Task" [ 841.238457] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.238912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9d2142ec-b2af-497a-ba3a-3c9cc241c9cc tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.006s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.249523] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224427, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.318197] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.318877] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.322971] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.425s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.323178] env[61947]: DEBUG nova.objects.instance [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lazy-loading 'resources' on Instance uuid cc1f84d4-0a28-48af-88ac-7d977db1cf34 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.445579] env[61947]: DEBUG oslo_vmware.api [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13896} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.445905] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.446143] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.446485] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.446650] env[61947]: INFO nova.compute.manager [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Took 1.13 seconds to destroy the instance on the hypervisor. [ 841.446914] env[61947]: DEBUG oslo.service.loopingcall [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.447155] env[61947]: DEBUG nova.compute.manager [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.447255] env[61947]: DEBUG nova.network.neutron [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.545644] env[61947]: DEBUG oslo_vmware.api [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Task: {'id': task-1224426, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134973} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.546510] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.546510] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.546510] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.546675] env[61947]: INFO nova.compute.manager [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Took 1.14 seconds to destroy the instance on the hypervisor. [ 841.546889] env[61947]: DEBUG oslo.service.loopingcall [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.547123] env[61947]: DEBUG nova.compute.manager [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.547223] env[61947]: DEBUG nova.network.neutron [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.668990] env[61947]: DEBUG nova.network.neutron [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.694397] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.758481] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224427, 'name': Rename_Task, 'duration_secs': 0.274409} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.759425] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 841.759425] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53985e65-d464-4459-9adf-5a63f2834de2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.775433] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 841.775433] env[61947]: value = "task-1224428" [ 841.775433] env[61947]: _type = "Task" [ 841.775433] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.794246] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224428, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.827265] env[61947]: DEBUG nova.compute.utils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.830999] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.830999] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.985298] env[61947]: DEBUG nova.policy [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd85ffd46682f45ef9302c6fd233ab4ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef226aa2daf94b9eaeb3b6b227292ff4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 842.176287] env[61947]: INFO nova.compute.manager [-] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Took 1.97 seconds to deallocate network for instance. [ 842.199135] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eeb8f5f-b42b-46c1-98e1-46ed2806e133 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.228630] env[61947]: DEBUG nova.compute.manager [req-fc9a0652-f0e6-42ef-917b-ff1751b536d9 req-21b1f3ff-0a02-4db7-8d2c-60f1779851c2 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Received event network-vif-deleted-1f213945-0710-459d-858f-c5e6ef01a82d {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.228849] env[61947]: INFO nova.compute.manager [req-fc9a0652-f0e6-42ef-917b-ff1751b536d9 req-21b1f3ff-0a02-4db7-8d2c-60f1779851c2 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Neutron deleted interface 1f213945-0710-459d-858f-c5e6ef01a82d; detaching it from the instance and deleting it from the info cache [ 842.229080] env[61947]: DEBUG nova.network.neutron [req-fc9a0652-f0e6-42ef-917b-ff1751b536d9 req-21b1f3ff-0a02-4db7-8d2c-60f1779851c2 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.231657] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "e0eebf59-c84e-4462-b280-d783a04525e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.231657] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.236341] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3f8a31-a4fd-4f17-8420-cfb76d262e4e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.274084] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b026f639-10a3-4002-952c-8cb16316a68f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.287548] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224428, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.291069] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165a336c-0184-46bb-93f8-87cbee76632b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.306912] env[61947]: DEBUG nova.compute.provider_tree [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.336855] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.535237] env[61947]: DEBUG nova.network.neutron [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.607497] env[61947]: DEBUG nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Received event network-vif-deleted-2a7b991a-9fca-4ed2-8096-6317d8bdc7f4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.607740] env[61947]: DEBUG nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-changed-1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.607912] env[61947]: DEBUG nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing instance network info cache due to event network-changed-1b562a73-9b70-4c31-9d83-f291e5b2cb76. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.608177] env[61947]: DEBUG oslo_concurrency.lockutils [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.608349] env[61947]: DEBUG oslo_concurrency.lockutils [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.608591] env[61947]: DEBUG nova.network.neutron [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing network info cache for port 1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.613842] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Successfully created port: c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.688097] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.741108] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 842.744738] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5030d069-ad9f-49ea-9baf-950119844495 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.758235] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a31225-260b-498a-b7c8-ec5c3b14ddb8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.792270] env[61947]: DEBUG nova.compute.manager [req-fc9a0652-f0e6-42ef-917b-ff1751b536d9 req-21b1f3ff-0a02-4db7-8d2c-60f1779851c2 service nova] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Detach interface failed, port_id=1f213945-0710-459d-858f-c5e6ef01a82d, reason: Instance f64b9f7f-89a9-4fb2-9c76-13cc591b9d51 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 842.796030] env[61947]: DEBUG oslo_vmware.api [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224428, 'name': PowerOnVM_Task, 'duration_secs': 0.937848} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.796030] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 842.796030] env[61947]: INFO nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Took 6.00 seconds to spawn the instance on the hypervisor. [ 842.796223] env[61947]: DEBUG nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.797299] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2506b14-34e2-4bd0-b9cf-3f35027b3821 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.810280] env[61947]: DEBUG nova.scheduler.client.report [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.865484] env[61947]: DEBUG nova.network.neutron [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.040957] env[61947]: INFO nova.compute.manager [-] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Took 1.49 seconds to deallocate network for instance. [ 843.266023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.315485] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.317602] env[61947]: INFO nova.compute.manager [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Took 32.12 seconds to build instance. [ 843.321705] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 26.233s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.321705] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.321705] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 843.321705] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.755s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.321705] env[61947]: INFO nova.compute.claims [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.326513] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b187a91a-f1ab-40a9-9c4b-bd777005e8c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.335950] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee92c630-01aa-42b7-9c24-e10c70c74a88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.353684] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.357115] env[61947]: INFO nova.scheduler.client.report [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Deleted allocations for instance cc1f84d4-0a28-48af-88ac-7d977db1cf34 [ 843.361955] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444b1d35-192c-4f3c-a9b8-b5b865797eda {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.368222] env[61947]: INFO nova.compute.manager [-] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Took 1.92 seconds to deallocate network for instance. [ 843.380435] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b81c1a-f24e-4178-8214-26a05b18a58b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.421111] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180999MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 843.421111] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.424425] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:48:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='9a82a007-fbf5-4f42-9bd3-1a8156f71854',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-569079550',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.424679] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.425101] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.425871] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.425871] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.425871] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.426085] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.426201] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.426388] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.427026] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.427026] env[61947]: DEBUG nova.virt.hardware [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.427577] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3a4e63-0ed0-495e-9c0e-bac98f061f18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.440060] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c093e3d-a21f-455c-b9a8-fb8c84a0c343 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.549539] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.602540] env[61947]: DEBUG nova.network.neutron [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updated VIF entry in instance network info cache for port 1b562a73-9b70-4c31-9d83-f291e5b2cb76. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.603017] env[61947]: DEBUG nova.network.neutron [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.824951] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8e502976-522b-40bc-bec0-67a3d56c7a3a tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.101s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.868911] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ad031410-b756-42b6-9dd6-86e57af48ad9 tempest-ServerAddressesNegativeTestJSON-69032506 tempest-ServerAddressesNegativeTestJSON-69032506-project-member] Lock "cc1f84d4-0a28-48af-88ac-7d977db1cf34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.081s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.881478] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.106353] env[61947]: DEBUG oslo_concurrency.lockutils [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.106353] env[61947]: DEBUG nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Received event network-vif-deleted-d1be0e35-dba1-4d37-b9ba-f1bcded1949a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.106353] env[61947]: INFO nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Neutron deleted interface d1be0e35-dba1-4d37-b9ba-f1bcded1949a; detaching it from the instance and deleting it from the info cache [ 844.106622] env[61947]: DEBUG nova.network.neutron [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.480850] env[61947]: DEBUG nova.compute.manager [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Stashing vm_state: active {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 844.494385] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Successfully updated port: c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.611893] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd127002-b1e6-4b01-9c28-d1d342f4e1d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.623103] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d497f3b-56dc-48a9-b899-02a589c0d43a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.657602] env[61947]: DEBUG nova.compute.manager [req-29ef3a7c-5f65-467b-a774-8d4b1da6da79 req-4841a631-d439-4b19-b7c9-d5ff375bdd0d service nova] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Detach interface failed, port_id=d1be0e35-dba1-4d37-b9ba-f1bcded1949a, reason: Instance f68a8537-5bae-4b50-b0ff-fa9b4a89bd40 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 844.663037] env[61947]: DEBUG nova.compute.manager [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Received event network-vif-plugged-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.663037] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Acquiring lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.663037] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.663377] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.663377] env[61947]: DEBUG nova.compute.manager [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] No waiting events found dispatching network-vif-plugged-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.663752] env[61947]: WARNING nova.compute.manager [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Received unexpected event network-vif-plugged-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 for instance with vm_state building and task_state spawning. [ 844.663752] env[61947]: DEBUG nova.compute.manager [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Received event network-changed-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.663841] env[61947]: DEBUG nova.compute.manager [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Refreshing instance network info cache due to event network-changed-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.664678] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.664678] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.664678] env[61947]: DEBUG nova.network.neutron [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Refreshing network info cache for port c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.698617] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947c984b-0e8d-4b47-851d-e965ea2b29c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.707102] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203ba8c0-9b6a-424c-800b-148a31f76c5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.738124] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01ef711-797e-4354-af87-8d2f86e95bb0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.747070] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39faab36-4a26-408a-ab31-cc78dafaf255 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.779068] env[61947]: DEBUG nova.compute.provider_tree [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.816568] env[61947]: DEBUG nova.compute.manager [None req-4b23fdcc-26ea-4909-a31b-5bec36ae4636 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.817484] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e14a95-f738-4158-a582-2590d6ec5a27 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.998672] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.004963] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.031800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.031800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.031800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "7d7f267a-3418-4b9b-9589-c915d5194f0d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.031800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.031800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.033048] env[61947]: INFO nova.compute.manager [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Terminating instance [ 845.034994] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "refresh_cache-7d7f267a-3418-4b9b-9589-c915d5194f0d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.035229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquired lock "refresh_cache-7d7f267a-3418-4b9b-9589-c915d5194f0d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.035418] env[61947]: DEBUG nova.network.neutron [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.227068] env[61947]: DEBUG nova.network.neutron [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.285024] env[61947]: DEBUG nova.scheduler.client.report [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.318022] env[61947]: DEBUG nova.network.neutron [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.329928] env[61947]: INFO nova.compute.manager [None req-4b23fdcc-26ea-4909-a31b-5bec36ae4636 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] instance snapshotting [ 845.329928] env[61947]: DEBUG nova.objects.instance [None req-4b23fdcc-26ea-4909-a31b-5bec36ae4636 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lazy-loading 'flavor' on Instance uuid 7d7f267a-3418-4b9b-9589-c915d5194f0d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.561415] env[61947]: DEBUG nova.network.neutron [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.668951] env[61947]: DEBUG nova.network.neutron [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.790784] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.790784] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.793736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.401s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.796120] env[61947]: INFO nova.compute.claims [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.819826] env[61947]: DEBUG oslo_concurrency.lockutils [req-b402a3a6-ddf4-4475-a9be-359a069592b0 req-673e2f35-aa80-4685-9f30-ee065b8ee396 service nova] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.822106] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.822106] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.835229] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422dcbc6-c8c4-4770-a556-0d62240ca501 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.855704] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf12372-2a2f-4bd9-ae03-4cd6a9533c39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.170542] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Releasing lock "refresh_cache-7d7f267a-3418-4b9b-9589-c915d5194f0d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.173196] env[61947]: DEBUG nova.compute.manager [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.173487] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 846.174497] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb21143-0969-4e2e-80e0-e3968cbbbba3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.184338] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 846.184338] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0b5c50c-b14c-4089-9cc6-1725bbec7a7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.194968] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 846.194968] env[61947]: value = "task-1224429" [ 846.194968] env[61947]: _type = "Task" [ 846.194968] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.203905] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.304510] env[61947]: DEBUG nova.compute.utils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.306273] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.306460] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.367685] env[61947]: DEBUG nova.compute.manager [None req-4b23fdcc-26ea-4909-a31b-5bec36ae4636 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance disappeared during snapshot {{(pid=61947) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 846.383471] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.392652] env[61947]: DEBUG nova.policy [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '897eb27f633a4a65bd500448ede60c1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7901e6bbfd1e4832a556d1a2b58d9850', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 846.555881] env[61947]: DEBUG nova.compute.manager [None req-4b23fdcc-26ea-4909-a31b-5bec36ae4636 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Found 0 images (rotation: 2) {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 846.703884] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224429, 'name': PowerOffVM_Task, 'duration_secs': 0.342449} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.704224] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 846.704433] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 846.704695] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32cb3acf-dd24-49b0-8749-5842914e37c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.735912] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 846.736190] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 846.736407] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Deleting the datastore file [datastore2] 7d7f267a-3418-4b9b-9589-c915d5194f0d {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.736690] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3266c831-bf57-48e4-aeb2-c9fa7c665214 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.748379] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for the task: (returnval){ [ 846.748379] env[61947]: value = "task-1224431" [ 846.748379] env[61947]: _type = "Task" [ 846.748379] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.776437] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.777367] env[61947]: DEBUG nova.network.neutron [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.809685] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.125221] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Successfully created port: 4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.165588] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d067a59e-3036-4c0b-b559-d8d9384e42a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.175206] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f3a6a4-5923-45a0-8ffb-ad302c065abc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.207808] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7929a8a5-1086-4b43-a42d-30b2e224f665 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.217549] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a63b41-13fb-455d-8254-d439fa3e5225 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.234633] env[61947]: DEBUG nova.compute.provider_tree [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.258825] env[61947]: DEBUG oslo_vmware.api [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Task: {'id': task-1224431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099345} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.259098] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.259379] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 847.259581] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 847.259664] env[61947]: INFO nova.compute.manager [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Took 1.09 seconds to destroy the instance on the hypervisor. [ 847.259913] env[61947]: DEBUG oslo.service.loopingcall [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.260126] env[61947]: DEBUG nova.compute.manager [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.260219] env[61947]: DEBUG nova.network.neutron [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.281380] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.281666] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance network_info: |[{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 847.282098] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:b8:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.290304] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Creating folder: Project (ef226aa2daf94b9eaeb3b6b227292ff4). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 847.290599] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-747a3756-0fd8-4e12-93a3-95a47f495e2f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.296994] env[61947]: DEBUG nova.network.neutron [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.304489] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Created folder: Project (ef226aa2daf94b9eaeb3b6b227292ff4) in parent group-v264556. [ 847.304826] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Creating folder: Instances. Parent ref: group-v264637. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 847.305228] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb30408f-141f-44ec-b5c7-1b5ac90df157 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.322273] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Created folder: Instances in parent group-v264637. [ 847.322534] env[61947]: DEBUG oslo.service.loopingcall [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.322736] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 847.322952] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9405c9e-fb31-4781-8793-a11a22903b7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.345938] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.345938] env[61947]: value = "task-1224434" [ 847.345938] env[61947]: _type = "Task" [ 847.345938] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.354351] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224434, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.737942] env[61947]: DEBUG nova.scheduler.client.report [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.799237] env[61947]: DEBUG nova.network.neutron [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.820868] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.853357] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.853616] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.853776] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.853969] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.854139] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.854292] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.854503] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.854666] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.854871] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.855058] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.855238] env[61947]: DEBUG nova.virt.hardware [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.856217] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984433f9-7a13-417f-be20-20d5c0ef995e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.869726] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaf9409-3410-412d-8507-203f9247cbba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.873755] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224434, 'name': CreateVM_Task, 'duration_secs': 0.382493} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.874276] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 847.875369] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.875559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.875917] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.876207] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8426a31b-8f65-4272-8598-579e32121dce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.891256] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 847.891256] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227ad9a-6ec3-37a1-2587-f54c53ad2a41" [ 847.891256] env[61947]: _type = "Task" [ 847.891256] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.899934] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227ad9a-6ec3-37a1-2587-f54c53ad2a41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.243920] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.245328] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.247241] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.583s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.248974] env[61947]: INFO nova.compute.claims [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.302438] env[61947]: INFO nova.compute.manager [-] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Took 1.04 seconds to deallocate network for instance. [ 848.404150] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227ad9a-6ec3-37a1-2587-f54c53ad2a41, 'name': SearchDatastore_Task, 'duration_secs': 0.0329} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.404522] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.404753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.405558] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.405558] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.405558] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.405558] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47c89494-1c76-4c43-8c58-02eed3f001ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.415909] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.416200] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 848.416963] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e5945f6-f2db-4926-b404-c31ee590f432 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.423832] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 848.423832] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52164f50-3d4f-d6ee-c954-fea9af88bfe7" [ 848.423832] env[61947]: _type = "Task" [ 848.423832] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.437773] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52164f50-3d4f-d6ee-c954-fea9af88bfe7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.753352] env[61947]: DEBUG nova.compute.utils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.761554] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.761768] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.812992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.851383] env[61947]: DEBUG nova.policy [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41d4a2d39ebc4ce0a0e76350f907e7df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58b83ea896bd4654aaa9e81f635a6a55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 848.935707] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52164f50-3d4f-d6ee-c954-fea9af88bfe7, 'name': SearchDatastore_Task, 'duration_secs': 0.010097} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.936623] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc44eff2-9d03-4c57-ade0-926a55c63d44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.943216] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 848.943216] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52123376-5ba3-c1b1-5416-ba83198f0619" [ 848.943216] env[61947]: _type = "Task" [ 848.943216] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.952577] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52123376-5ba3-c1b1-5416-ba83198f0619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.998103] env[61947]: DEBUG nova.compute.manager [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Received event network-vif-plugged-4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.998395] env[61947]: DEBUG oslo_concurrency.lockutils [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.998684] env[61947]: DEBUG oslo_concurrency.lockutils [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.999299] env[61947]: DEBUG oslo_concurrency.lockutils [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.999299] env[61947]: DEBUG nova.compute.manager [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] No waiting events found dispatching network-vif-plugged-4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.999504] env[61947]: WARNING nova.compute.manager [req-00d7b444-e632-4d8b-a202-306996880e68 req-1d3c1507-3a20-4645-b3a9-2b08b66eff58 service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Received unexpected event network-vif-plugged-4024dcff-f052-4a2e-9fab-c3d722069bbe for instance with vm_state building and task_state spawning. [ 849.138208] env[61947]: DEBUG nova.compute.manager [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.139546] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3b875d-945a-4e0a-a7bb-0c0df7bddd24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.171234] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Successfully updated port: 4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.263112] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.318249] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Successfully created port: 8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.455085] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52123376-5ba3-c1b1-5416-ba83198f0619, 'name': SearchDatastore_Task, 'duration_secs': 0.009873} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.457605] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.457886] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.458380] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfcd69d0-b081-43c9-a676-d08b526ea3ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.466412] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 849.466412] env[61947]: value = "task-1224435" [ 849.466412] env[61947]: _type = "Task" [ 849.466412] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.475877] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.547024] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fd3187-5c30-4c71-908a-3f5147df5156 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.556722] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98692d8f-7f5e-4d24-b1d6-cde94bf27dd2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.605896] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be694a1c-0928-477c-9566-c5a57b3da940 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.615201] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39429644-c690-4972-8ef3-909767022955 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.631486] env[61947]: DEBUG nova.compute.provider_tree [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.650769] env[61947]: INFO nova.compute.manager [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] instance snapshotting [ 849.651424] env[61947]: DEBUG nova.objects.instance [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.672323] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.672476] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.672640] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.978938] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224435, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.136307] env[61947]: DEBUG nova.scheduler.client.report [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.158187] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a730f0-760a-46e2-90ed-9536564338d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.184405] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94656d75-3f02-4c55-a9d3-36e16ee794a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.220127] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.276782] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.306875] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.307191] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.307399] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.307631] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.307821] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.308015] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.308368] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.308584] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.308724] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.308933] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.309168] env[61947]: DEBUG nova.virt.hardware [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.310176] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21ef783-99c6-4e94-a2cb-fd6922bf7e78 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.318656] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1faa9fb1-7111-449c-9bfd-a230fc322668 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.378876] env[61947]: DEBUG nova.network.neutron [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updating instance_info_cache with network_info: [{"id": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "address": "fa:16:3e:91:65:29", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4024dcff-f0", "ovs_interfaceid": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.478531] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224435, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543912} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.478869] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 850.479146] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.479459] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a480d2e4-a0a9-4061-aaf9-bac9b1c57e14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.487058] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 850.487058] env[61947]: value = "task-1224436" [ 850.487058] env[61947]: _type = "Task" [ 850.487058] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.499534] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.645296] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.645296] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.647674] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.644s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.650469] env[61947]: INFO nova.compute.claims [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.697738] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 850.698350] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cf512647-ccb9-4fd8-917e-e744b7750834 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.708174] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 850.708174] env[61947]: value = "task-1224437" [ 850.708174] env[61947]: _type = "Task" [ 850.708174] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.720310] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224437, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.885199] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.885199] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Instance network_info: |[{"id": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "address": "fa:16:3e:91:65:29", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4024dcff-f0", "ovs_interfaceid": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.885199] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:65:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4024dcff-f052-4a2e-9fab-c3d722069bbe', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.892850] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating folder: Project (7901e6bbfd1e4832a556d1a2b58d9850). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.893334] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bd7ec49-8572-4e4b-a63d-116a5d2ba259 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.909019] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created folder: Project (7901e6bbfd1e4832a556d1a2b58d9850) in parent group-v264556. [ 850.909019] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating folder: Instances. Parent ref: group-v264640. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.909019] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f081abe-6261-49ec-a70f-167d94b35da5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.920984] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created folder: Instances in parent group-v264640. [ 850.920984] env[61947]: DEBUG oslo.service.loopingcall [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.920984] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 850.921183] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5918613-dd0d-4345-8ae7-67a023e0fedb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.943906] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.943906] env[61947]: value = "task-1224440" [ 850.943906] env[61947]: _type = "Task" [ 850.943906] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.954197] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224440, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.999240] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063932} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.999557] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.000374] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c991c39d-95ad-482a-90d7-71f7c8500a3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.027218] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.027563] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f9be7dc-ebdf-4660-a762-7ee75fae5eb6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.050286] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 851.050286] env[61947]: value = "task-1224441" [ 851.050286] env[61947]: _type = "Task" [ 851.050286] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.057750] env[61947]: DEBUG nova.compute.manager [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Received event network-changed-4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.057835] env[61947]: DEBUG nova.compute.manager [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Refreshing instance network info cache due to event network-changed-4024dcff-f052-4a2e-9fab-c3d722069bbe. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 851.059458] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Acquiring lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.059458] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Acquired lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.059458] env[61947]: DEBUG nova.network.neutron [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Refreshing network info cache for port 4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.061406] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Successfully updated port: 8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.066137] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224441, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.157848] env[61947]: DEBUG nova.compute.utils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.159714] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.163021] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.221116] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224437, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.234958] env[61947]: DEBUG nova.policy [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd184f9216c164f54ae07005cb607109d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9ffe92d54eb4e7d91e95f68f195c49e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 851.455357] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224440, 'name': CreateVM_Task, 'duration_secs': 0.321702} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.455929] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.456589] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.456589] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.456753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.457035] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d87570b-8818-4fcc-9d77-181fc14f261a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.463023] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 851.463023] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52786537-bdf4-ea76-170f-631dce6f36f1" [ 851.463023] env[61947]: _type = "Task" [ 851.463023] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.470351] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52786537-bdf4-ea76-170f-631dce6f36f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.562549] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224441, 'name': ReconfigVM_Task, 'duration_secs': 0.302538} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.562549] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.562549] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2645ee75-2acb-496e-94cb-1e3685239df1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.572141] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.572141] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.572141] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.572141] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 851.572141] env[61947]: value = "task-1224442" [ 851.572141] env[61947]: _type = "Task" [ 851.572141] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.581988] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224442, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.664728] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.680156] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Successfully created port: 1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.727451] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224437, 'name': CreateSnapshot_Task, 'duration_secs': 0.541468} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.730111] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 851.731119] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b28309b-aab1-415c-acea-dbf9ec8d4a52 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.977830] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52786537-bdf4-ea76-170f-631dce6f36f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009993} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.980469] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.980710] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.980945] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.981110] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.981295] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.981765] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b5baccd-2fa9-4253-ac67-349f360a6b88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.992133] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.992133] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 851.998767] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b59ab01-4cb3-40c0-bc37-e5f076dd9a51 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.999925] env[61947]: DEBUG nova.network.neutron [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updated VIF entry in instance network info cache for port 4024dcff-f052-4a2e-9fab-c3d722069bbe. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.000328] env[61947]: DEBUG nova.network.neutron [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updating instance_info_cache with network_info: [{"id": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "address": "fa:16:3e:91:65:29", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4024dcff-f0", "ovs_interfaceid": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.005187] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 852.005187] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a9e724-4b7f-90dd-34b1-e0784aa7b73d" [ 852.005187] env[61947]: _type = "Task" [ 852.005187] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.017049] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a9e724-4b7f-90dd-34b1-e0784aa7b73d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.065839] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05022db9-5102-4b28-8a79-120e1267955f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.077054] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0da4b14-d64e-48c5-bc99-c5a23c632901 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.088659] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224442, 'name': Rename_Task, 'duration_secs': 0.148459} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.114015] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 852.114920] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3718bec6-9178-4e1a-a043-57d5c269ce58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.118286] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ec1066-4699-4cbc-a7bb-ce2b98bb1a8a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.128852] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.131789] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e70f4c-4dad-4855-9661-880c5bf2695c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.136960] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 852.136960] env[61947]: value = "task-1224443" [ 852.136960] env[61947]: _type = "Task" [ 852.136960] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.149193] env[61947]: DEBUG nova.compute.provider_tree [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.157718] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224443, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.249548] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 852.249881] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9a0540d8-7d05-4e91-86aa-4cc8eeb2f363 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.259664] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 852.259664] env[61947]: value = "task-1224444" [ 852.259664] env[61947]: _type = "Task" [ 852.259664] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.271343] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224444, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.502702] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Releasing lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.503033] env[61947]: DEBUG nova.compute.manager [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Received event network-vif-plugged-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.503264] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Acquiring lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.503494] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.503692] env[61947]: DEBUG oslo_concurrency.lockutils [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.503835] env[61947]: DEBUG nova.compute.manager [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] No waiting events found dispatching network-vif-plugged-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.504016] env[61947]: WARNING nova.compute.manager [req-480801c5-cc3c-41a8-93db-e3446feb1430 req-3216af39-c8f4-4b1d-a592-5b09d31b0dcd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Received unexpected event network-vif-plugged-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a for instance with vm_state building and task_state spawning. [ 852.516701] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a9e724-4b7f-90dd-34b1-e0784aa7b73d, 'name': SearchDatastore_Task, 'duration_secs': 0.009139} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.517521] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be3963b1-62d2-49ac-91e8-91bfb187ad29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.524696] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 852.524696] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525513ae-c972-40bf-0ee2-c5ef87565f93" [ 852.524696] env[61947]: _type = "Task" [ 852.524696] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.535218] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525513ae-c972-40bf-0ee2-c5ef87565f93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.619148] env[61947]: DEBUG nova.network.neutron [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Updating instance_info_cache with network_info: [{"id": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "address": "fa:16:3e:2e:36:77", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa0162f-f4", "ovs_interfaceid": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.647494] env[61947]: DEBUG oslo_vmware.api [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224443, 'name': PowerOnVM_Task, 'duration_secs': 0.474125} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.647783] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 852.647993] env[61947]: INFO nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Took 9.29 seconds to spawn the instance on the hypervisor. [ 852.648824] env[61947]: DEBUG nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.649721] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4601df-f5e8-46ee-83d9-0dfe0cccb19b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.653287] env[61947]: DEBUG nova.scheduler.client.report [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.678882] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.718419] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.718962] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.719335] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.719857] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.720243] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.721373] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.721373] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.721373] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.721816] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.722120] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.722435] env[61947]: DEBUG nova.virt.hardware [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.723729] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b407e02-10ea-4651-917e-7ee6ae712521 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.734070] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05218c8-a1f6-42cd-a7af-30b79976fec3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.773039] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224444, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.837912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.837912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.035841] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525513ae-c972-40bf-0ee2-c5ef87565f93, 'name': SearchDatastore_Task, 'duration_secs': 0.010283} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.036115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.036630] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] bd84b03e-ef6d-41e0-b48c-b8e8fc425a36/bd84b03e-ef6d-41e0-b48c-b8e8fc425a36.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.036727] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b81c037e-def4-4858-81eb-d614d2a747ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.043827] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 853.043827] env[61947]: value = "task-1224445" [ 853.043827] env[61947]: _type = "Task" [ 853.043827] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.052457] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.088679] env[61947]: DEBUG nova.compute.manager [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Received event network-changed-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.088944] env[61947]: DEBUG nova.compute.manager [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Refreshing instance network info cache due to event network-changed-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 853.089182] env[61947]: DEBUG oslo_concurrency.lockutils [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] Acquiring lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.122390] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.122754] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Instance network_info: |[{"id": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "address": "fa:16:3e:2e:36:77", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa0162f-f4", "ovs_interfaceid": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.123082] env[61947]: DEBUG oslo_concurrency.lockutils [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] Acquired lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.123275] env[61947]: DEBUG nova.network.neutron [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Refreshing network info cache for port 8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.124524] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:36:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fa0162f-f4dc-4407-9a88-0f6c75ffd12a', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.132560] env[61947]: DEBUG oslo.service.loopingcall [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.133500] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.133733] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d10d0088-83f2-48b4-9eda-6663ac986ea4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.156271] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.156271] env[61947]: value = "task-1224446" [ 853.156271] env[61947]: _type = "Task" [ 853.156271] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.161446] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.162077] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.164988] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.882s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.166851] env[61947]: INFO nova.compute.claims [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.181389] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224446, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.182047] env[61947]: INFO nova.compute.manager [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Took 40.49 seconds to build instance. [ 853.272204] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224444, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.339702] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.555302] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471291} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.556725] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] bd84b03e-ef6d-41e0-b48c-b8e8fc425a36/bd84b03e-ef6d-41e0-b48c-b8e8fc425a36.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 853.556725] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.556725] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbf18011-48df-4058-8d57-57e8a215b033 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.566031] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 853.566031] env[61947]: value = "task-1224447" [ 853.566031] env[61947]: _type = "Task" [ 853.566031] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.575541] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224447, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.666952] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224446, 'name': CreateVM_Task, 'duration_secs': 0.459204} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.667232] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 853.667938] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.668086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.668433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.668725] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b51066fb-de7b-4737-9bf2-80dfe96ac19d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.679415] env[61947]: DEBUG nova.compute.utils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.688022] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 853.688022] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b657e-a684-9e57-6b25-99c3e1a2e69c" [ 853.688022] env[61947]: _type = "Task" [ 853.688022] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.688022] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.688022] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.689062] env[61947]: DEBUG oslo_concurrency.lockutils [None req-af15136d-dee3-4d64-87ff-477c58c740cb tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.867s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.701490] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b657e-a684-9e57-6b25-99c3e1a2e69c, 'name': SearchDatastore_Task, 'duration_secs': 0.010257} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.702458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.702697] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.703009] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.703177] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.703362] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.703896] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fcfcc414-e0ad-4113-8f2e-c38855fe5e5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.720035] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.720252] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 853.721040] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02997526-9235-4a09-acb3-af8005258f58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.727871] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 853.727871] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522c0c19-a30c-1259-e323-3e7637a81568" [ 853.727871] env[61947]: _type = "Task" [ 853.727871] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.737873] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522c0c19-a30c-1259-e323-3e7637a81568, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.760057] env[61947]: DEBUG nova.policy [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7cafa3cd6ef416b8447dbc1ccfeaa02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3370eb1527c41dcb1037f3ef73705fc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 853.772711] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224444, 'name': CloneVM_Task, 'duration_secs': 1.443599} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.773846] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created linked-clone VM from snapshot [ 853.776426] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e51ae9-ad44-4764-b0ea-09f8c0efc995 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.787719] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploading image 1cb3a519-7891-4eeb-9474-821506e882fb {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 853.820130] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 853.820130] env[61947]: value = "vm-264644" [ 853.820130] env[61947]: _type = "VirtualMachine" [ 853.820130] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 853.820893] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d9b1f91e-bc89-4cf7-b754-bf2d41210c71 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.829732] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease: (returnval){ [ 853.829732] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a89ce-14a5-4fdc-e893-48291475744c" [ 853.829732] env[61947]: _type = "HttpNfcLease" [ 853.829732] env[61947]: } obtained for exporting VM: (result){ [ 853.829732] env[61947]: value = "vm-264644" [ 853.829732] env[61947]: _type = "VirtualMachine" [ 853.829732] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 853.830038] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the lease: (returnval){ [ 853.830038] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a89ce-14a5-4fdc-e893-48291475744c" [ 853.830038] env[61947]: _type = "HttpNfcLease" [ 853.830038] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 853.837820] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 853.837820] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a89ce-14a5-4fdc-e893-48291475744c" [ 853.837820] env[61947]: _type = "HttpNfcLease" [ 853.837820] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 853.859772] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.077118] env[61947]: DEBUG nova.network.neutron [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Updated VIF entry in instance network info cache for port 8fa0162f-f4dc-4407-9a88-0f6c75ffd12a. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.077118] env[61947]: DEBUG nova.network.neutron [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Updating instance_info_cache with network_info: [{"id": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "address": "fa:16:3e:2e:36:77", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa0162f-f4", "ovs_interfaceid": "8fa0162f-f4dc-4407-9a88-0f6c75ffd12a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.085580] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Successfully updated port: 1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.093968] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224447, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122238} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.095492] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.096548] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026e3848-e5e6-4cd6-886f-d8e926cc7d4d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.125031] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] bd84b03e-ef6d-41e0-b48c-b8e8fc425a36/bd84b03e-ef6d-41e0-b48c-b8e8fc425a36.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.125870] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-691daff8-fe9a-4e86-9826-b17fa253d661 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.148487] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 854.148487] env[61947]: value = "task-1224449" [ 854.148487] env[61947]: _type = "Task" [ 854.148487] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.159840] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224449, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.191263] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.211880] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Successfully created port: d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.242780] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522c0c19-a30c-1259-e323-3e7637a81568, 'name': SearchDatastore_Task, 'duration_secs': 0.025153} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.246244] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a251587-5d26-4848-8e60-2209f318ec8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.253871] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 854.253871] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e88bd2-706d-7546-1be1-2d442fa9585d" [ 854.253871] env[61947]: _type = "Task" [ 854.253871] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.263305] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e88bd2-706d-7546-1be1-2d442fa9585d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.338763] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 854.338763] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a89ce-14a5-4fdc-e893-48291475744c" [ 854.338763] env[61947]: _type = "HttpNfcLease" [ 854.338763] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 854.341797] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 854.341797] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a89ce-14a5-4fdc-e893-48291475744c" [ 854.341797] env[61947]: _type = "HttpNfcLease" [ 854.341797] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 854.342856] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad69402a-744f-47f1-84b4-dc2df7282d37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.354022] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 854.354022] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 854.469275] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3211f42c-4ba5-439b-9a5c-c88a7bedfb1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.586067] env[61947]: DEBUG oslo_concurrency.lockutils [req-f1fb08fa-404c-47b2-be97-74c3e7d7ebac req-5017a1c8-f173-4291-90e7-12568e24d6bd service nova] Releasing lock "refresh_cache-9bc18645-c837-4dab-8e05-fbcf0d40812f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.591531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.591654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.591786] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.618579] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed579123-ef8a-475d-ba9c-d6fe2f4b91b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.629876] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9931e8f-9604-4076-be83-18eefcc0d0bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.669143] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e5d29d-be97-4f0c-923f-04d9c116eb37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.678345] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224449, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.683536] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4caec7-08a9-4a45-b0fa-6a1d55a44c79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.702351] env[61947]: DEBUG nova.compute.provider_tree [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.765211] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e88bd2-706d-7546-1be1-2d442fa9585d, 'name': SearchDatastore_Task, 'duration_secs': 0.009991} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.765549] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.765939] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9bc18645-c837-4dab-8e05-fbcf0d40812f/9bc18645-c837-4dab-8e05-fbcf0d40812f.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 854.766338] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-796c01cd-5d46-49f7-8b98-8d5c10c806e0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.776942] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 854.776942] env[61947]: value = "task-1224450" [ 854.776942] env[61947]: _type = "Task" [ 854.776942] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.787087] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.941304] env[61947]: DEBUG nova.compute.manager [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Stashing vm_state: active {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 855.143612] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.178733] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224449, 'name': ReconfigVM_Task, 'duration_secs': 0.969572} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.179381] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfigured VM instance instance-00000040 to attach disk [datastore2] bd84b03e-ef6d-41e0-b48c-b8e8fc425a36/bd84b03e-ef6d-41e0-b48c-b8e8fc425a36.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.181044] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15368110-143e-4b7d-b1b5-dc164b63b949 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.190341] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 855.190341] env[61947]: value = "task-1224451" [ 855.190341] env[61947]: _type = "Task" [ 855.190341] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.202348] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224451, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.205727] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.209294] env[61947]: DEBUG nova.scheduler.client.report [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.218925] env[61947]: DEBUG nova.compute.manager [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Received event network-vif-plugged-1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.219222] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.219708] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.221077] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.221288] env[61947]: DEBUG nova.compute.manager [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] No waiting events found dispatching network-vif-plugged-1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.221467] env[61947]: WARNING nova.compute.manager [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Received unexpected event network-vif-plugged-1a026df8-fa69-439e-805b-5dad5c772246 for instance with vm_state building and task_state spawning. [ 855.221641] env[61947]: DEBUG nova.compute.manager [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Received event network-changed-1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.221824] env[61947]: DEBUG nova.compute.manager [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Refreshing instance network info cache due to event network-changed-1a026df8-fa69-439e-805b-5dad5c772246. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 855.222030] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Acquiring lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.243251] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.243589] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.244742] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.244742] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.244742] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.244742] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.244742] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.245066] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.245066] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.245266] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.245642] env[61947]: DEBUG nova.virt.hardware [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.247182] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ae9beb-312a-465c-a3cc-d89c9ad173c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.259847] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f6c7ad-5949-4243-a1a4-59f8d8a0d6ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.287436] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224450, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.375420] env[61947]: DEBUG nova.network.neutron [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updating instance_info_cache with network_info: [{"id": "1a026df8-fa69-439e-805b-5dad5c772246", "address": "fa:16:3e:5d:8b:d0", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a026df8-fa", "ovs_interfaceid": "1a026df8-fa69-439e-805b-5dad5c772246", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.468978] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.702479] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224451, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.715605] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.716272] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.719289] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.354s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.719794] env[61947]: DEBUG nova.objects.instance [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lazy-loading 'resources' on Instance uuid b7d84a56-edb6-4d1e-b7e9-294078be79d8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.788365] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654047} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.789072] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9bc18645-c837-4dab-8e05-fbcf0d40812f/9bc18645-c837-4dab-8e05-fbcf0d40812f.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 855.789267] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.789536] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a2f442c-3f54-47fd-980b-d00cccc9605e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.798541] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 855.798541] env[61947]: value = "task-1224452" [ 855.798541] env[61947]: _type = "Task" [ 855.798541] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.809802] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224452, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.880986] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.880986] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Instance network_info: |[{"id": "1a026df8-fa69-439e-805b-5dad5c772246", "address": "fa:16:3e:5d:8b:d0", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a026df8-fa", "ovs_interfaceid": "1a026df8-fa69-439e-805b-5dad5c772246", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.880986] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Acquired lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.880986] env[61947]: DEBUG nova.network.neutron [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Refreshing network info cache for port 1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.882360] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:8b:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '517421c3-bea0-419c-ab0b-987815e5d160', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a026df8-fa69-439e-805b-5dad5c772246', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.893367] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Creating folder: Project (c9ffe92d54eb4e7d91e95f68f195c49e). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 855.895151] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6caf3d13-9031-4310-9ac7-22350a4ab27d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.910453] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Created folder: Project (c9ffe92d54eb4e7d91e95f68f195c49e) in parent group-v264556. [ 855.910800] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Creating folder: Instances. Parent ref: group-v264646. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 855.911206] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03c522d6-9573-4fd8-92b7-079f9a4c3c5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.925319] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Created folder: Instances in parent group-v264646. [ 855.925738] env[61947]: DEBUG oslo.service.loopingcall [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.925980] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 855.926251] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e017ea63-2ea9-4a62-b861-796df3686628 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.957143] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.957143] env[61947]: value = "task-1224455" [ 855.957143] env[61947]: _type = "Task" [ 855.957143] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.968671] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224455, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.056481] env[61947]: DEBUG nova.compute.manager [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Received event network-vif-plugged-d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.056481] env[61947]: DEBUG oslo_concurrency.lockutils [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] Acquiring lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.056481] env[61947]: DEBUG oslo_concurrency.lockutils [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.056481] env[61947]: DEBUG oslo_concurrency.lockutils [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.057325] env[61947]: DEBUG nova.compute.manager [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] No waiting events found dispatching network-vif-plugged-d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.057688] env[61947]: WARNING nova.compute.manager [req-17157ec5-9c13-4c6b-a5a9-036128721c47 req-c2c68aa3-e9d8-4ce7-96d0-298b0df27459 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Received unexpected event network-vif-plugged-d47b903b-da5e-4363-9b20-4d4238cdc3e2 for instance with vm_state building and task_state spawning. [ 856.175446] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Successfully updated port: d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.204655] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224451, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.222783] env[61947]: DEBUG nova.compute.utils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.225133] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.225406] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.300844] env[61947]: DEBUG nova.policy [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5be984ccbc9486b9f15cb279178f027', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '298bd92edec8447a9377fac94e783f16', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 856.315632] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224452, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079586} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.316825] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.317131] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cff243-888d-4f28-aa1e-c35fa8677b27 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.343151] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 9bc18645-c837-4dab-8e05-fbcf0d40812f/9bc18645-c837-4dab-8e05-fbcf0d40812f.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.351179] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef7d4d55-ea2b-4f87-aae3-fe8033bb2e32 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.374532] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 856.374532] env[61947]: value = "task-1224456" [ 856.374532] env[61947]: _type = "Task" [ 856.374532] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.390019] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224456, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.468658] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224455, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.649069] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2906885-da42-484b-ab3b-1eeeda35dc25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.657719] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab4d300-4d53-4a44-8f15-cfabfe5c6514 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.695107] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.695385] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquired lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.695639] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.705768] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbaaab5-9b3b-423c-9897-209afa1758c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.713189] env[61947]: DEBUG nova.network.neutron [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updated VIF entry in instance network info cache for port 1a026df8-fa69-439e-805b-5dad5c772246. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 856.713189] env[61947]: DEBUG nova.network.neutron [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updating instance_info_cache with network_info: [{"id": "1a026df8-fa69-439e-805b-5dad5c772246", "address": "fa:16:3e:5d:8b:d0", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a026df8-fa", "ovs_interfaceid": "1a026df8-fa69-439e-805b-5dad5c772246", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.716584] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Successfully created port: 7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.726650] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224451, 'name': Rename_Task, 'duration_secs': 1.234098} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.727176] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.730013] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.734894] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae795e6-516c-4a0a-b393-d1ce32826a38 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.740063] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fda6bf7-f856-45a6-a5ff-cb54883e72fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.755608] env[61947]: DEBUG nova.compute.provider_tree [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.759182] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 856.759182] env[61947]: value = "task-1224457" [ 856.759182] env[61947]: _type = "Task" [ 856.759182] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.769568] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.886413] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224456, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.968491] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224455, 'name': CreateVM_Task, 'duration_secs': 0.573754} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.968825] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.969622] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.969796] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.970159] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.970476] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37d5899a-09b5-44b7-b72f-2dcbe1a9e7f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.976046] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 856.976046] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52358f3a-8757-606d-4eab-8d9dcaa9cf69" [ 856.976046] env[61947]: _type = "Task" [ 856.976046] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.985822] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52358f3a-8757-606d-4eab-8d9dcaa9cf69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.221260] env[61947]: DEBUG oslo_concurrency.lockutils [req-b78ca8d5-255a-40bb-94f0-e69c90eaa87f req-80e1bc9f-a482-4547-a7d4-c217821ad7be service nova] Releasing lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.258019] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.260762] env[61947]: DEBUG nova.scheduler.client.report [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.274154] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224457, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.386068] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224456, 'name': ReconfigVM_Task, 'duration_secs': 0.610853} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.386210] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 9bc18645-c837-4dab-8e05-fbcf0d40812f/9bc18645-c837-4dab-8e05-fbcf0d40812f.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.386837] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-385af14f-2787-4856-ac0b-1d21c56b1da7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.394677] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 857.394677] env[61947]: value = "task-1224458" [ 857.394677] env[61947]: _type = "Task" [ 857.394677] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.403977] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224458, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.458513] env[61947]: DEBUG nova.network.neutron [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Updating instance_info_cache with network_info: [{"id": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "address": "fa:16:3e:39:31:a6", "network": {"id": "f382b00f-7710-430a-9107-7c37dcd65382", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1545289290-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3370eb1527c41dcb1037f3ef73705fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47b903b-da", "ovs_interfaceid": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.489628] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52358f3a-8757-606d-4eab-8d9dcaa9cf69, 'name': SearchDatastore_Task, 'duration_secs': 0.011502} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.490412] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.491045] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.491045] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.491150] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.491324] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.491605] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0445536-9d25-4757-8e11-dddf2a66a144 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.503657] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.503914] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.504742] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b829a6-6620-4029-8245-497eafcc6619 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.512499] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 857.512499] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521793f9-1d58-f79a-5f00-fe9db5e87b56" [ 857.512499] env[61947]: _type = "Task" [ 857.512499] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.522473] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521793f9-1d58-f79a-5f00-fe9db5e87b56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.746656] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.768164] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.049s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.774756] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.196s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.774756] env[61947]: DEBUG nova.objects.instance [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lazy-loading 'resources' on Instance uuid 74e03575-297e-4e08-9236-98d8be80b546 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.777717] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.777717] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.777967] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.778046] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.778167] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.778457] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.778798] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.778798] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.778915] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.779075] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.779487] env[61947]: DEBUG nova.virt.hardware [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.780405] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4ad8cb-c6ae-442b-bab6-be68bee43d4d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.790619] env[61947]: DEBUG oslo_vmware.api [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224457, 'name': PowerOnVM_Task, 'duration_secs': 0.879487} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.792995] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.793247] env[61947]: INFO nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Took 9.97 seconds to spawn the instance on the hypervisor. [ 857.793436] env[61947]: DEBUG nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.794492] env[61947]: INFO nova.scheduler.client.report [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Deleted allocations for instance b7d84a56-edb6-4d1e-b7e9-294078be79d8 [ 857.796078] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415ac717-9ae9-43d5-afd7-b0fcb41916d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.800903] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b36bcc-6b3a-4436-900e-8c830166d6b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.905801] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224458, 'name': Rename_Task, 'duration_secs': 0.234618} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.906123] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 857.906403] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c668322b-9384-484a-bb22-38ed5c09b784 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.913680] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 857.913680] env[61947]: value = "task-1224459" [ 857.913680] env[61947]: _type = "Task" [ 857.913680] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.922511] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.961667] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Releasing lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.962231] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Instance network_info: |[{"id": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "address": "fa:16:3e:39:31:a6", "network": {"id": "f382b00f-7710-430a-9107-7c37dcd65382", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1545289290-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3370eb1527c41dcb1037f3ef73705fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47b903b-da", "ovs_interfaceid": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.962743] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:31:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f54f7284-8f7d-47ee-839d-2143062cfe44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd47b903b-da5e-4363-9b20-4d4238cdc3e2', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.971603] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Creating folder: Project (e3370eb1527c41dcb1037f3ef73705fc). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.972279] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5310da0b-48c8-421e-b75a-e388c57a4ae8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.986337] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Created folder: Project (e3370eb1527c41dcb1037f3ef73705fc) in parent group-v264556. [ 857.986585] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Creating folder: Instances. Parent ref: group-v264649. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.986912] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1666239-1376-40b1-9c2a-9e59eed69c13 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.000527] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Created folder: Instances in parent group-v264649. [ 858.000818] env[61947]: DEBUG oslo.service.loopingcall [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.001085] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.001324] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f331962-f00a-4c95-8348-6678467032e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.026078] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521793f9-1d58-f79a-5f00-fe9db5e87b56, 'name': SearchDatastore_Task, 'duration_secs': 0.012667} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.028054] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.028054] env[61947]: value = "task-1224462" [ 858.028054] env[61947]: _type = "Task" [ 858.028054] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.028295] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36c87573-cbc6-420a-a201-1a67c9195311 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.039725] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224462, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.040840] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 858.040840] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52018d7f-477e-a8e6-ce0f-f9e9bb4ae4a8" [ 858.040840] env[61947]: _type = "Task" [ 858.040840] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.050907] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52018d7f-477e-a8e6-ce0f-f9e9bb4ae4a8, 'name': SearchDatastore_Task, 'duration_secs': 0.010644} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.051211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.051500] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 8c342d64-75f6-4b7d-829d-889eb3b6122f/8c342d64-75f6-4b7d-829d-889eb3b6122f.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.051774] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36dcd192-b246-41c3-a73e-7ca7bc9f651a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.060744] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 858.060744] env[61947]: value = "task-1224463" [ 858.060744] env[61947]: _type = "Task" [ 858.060744] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.069912] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.093542] env[61947]: DEBUG nova.compute.manager [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Received event network-changed-d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.093779] env[61947]: DEBUG nova.compute.manager [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Refreshing instance network info cache due to event network-changed-d47b903b-da5e-4363-9b20-4d4238cdc3e2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.093931] env[61947]: DEBUG oslo_concurrency.lockutils [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] Acquiring lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.094166] env[61947]: DEBUG oslo_concurrency.lockutils [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] Acquired lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.094516] env[61947]: DEBUG nova.network.neutron [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Refreshing network info cache for port d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.317709] env[61947]: DEBUG oslo_concurrency.lockutils [None req-812bdde0-e1e2-4ce2-aeae-2b33412f71b7 tempest-ServersTestJSON-1122712346 tempest-ServersTestJSON-1122712346-project-member] Lock "b7d84a56-edb6-4d1e-b7e9-294078be79d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.942s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.343071] env[61947]: INFO nova.compute.manager [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Took 34.81 seconds to build instance. [ 858.433766] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224459, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.436969] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Successfully updated port: 7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 858.545473] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224462, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.572601] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224463, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.713965] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d7c0ff-ad6c-4117-8b1d-530622f9063e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.723315] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4e035b-ce69-407d-998a-c776f8b8d326 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.763909] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a3d51b-7ef4-4f14-86df-a738c325607e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.773475] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3841ae-76d4-45c4-bf16-9525b188b281 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.790515] env[61947]: DEBUG nova.compute.provider_tree [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.846499] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f9d080ff-2b42-4e5c-b480-9d3abf2b285f tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.805s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.929111] env[61947]: DEBUG oslo_vmware.api [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224459, 'name': PowerOnVM_Task, 'duration_secs': 0.734091} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.932021] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 858.932021] env[61947]: INFO nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Took 8.65 seconds to spawn the instance on the hypervisor. [ 858.932021] env[61947]: DEBUG nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.932021] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4097688c-3c13-4a7c-9b43-9cd339b598a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.943502] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.943638] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquired lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.943777] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.042770] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224462, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.077187] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224463, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.104125] env[61947]: DEBUG nova.network.neutron [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Updated VIF entry in instance network info cache for port d47b903b-da5e-4363-9b20-4d4238cdc3e2. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.104224] env[61947]: DEBUG nova.network.neutron [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Updating instance_info_cache with network_info: [{"id": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "address": "fa:16:3e:39:31:a6", "network": {"id": "f382b00f-7710-430a-9107-7c37dcd65382", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1545289290-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3370eb1527c41dcb1037f3ef73705fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47b903b-da", "ovs_interfaceid": "d47b903b-da5e-4363-9b20-4d4238cdc3e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.295742] env[61947]: DEBUG nova.scheduler.client.report [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.451812] env[61947]: INFO nova.compute.manager [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Took 34.09 seconds to build instance. [ 859.487720] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.549486] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224462, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.576046] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224463, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.607748] env[61947]: DEBUG oslo_concurrency.lockutils [req-8749ee0c-436d-4964-a751-f3fadcad4434 req-a5a89ada-041f-45b0-ab6e-2845a4ae1f93 service nova] Releasing lock "refresh_cache-4f0b9636-4793-4b1c-8b2b-7f31a15669bb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.673054] env[61947]: DEBUG nova.network.neutron [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Updating instance_info_cache with network_info: [{"id": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "address": "fa:16:3e:aa:64:bf", "network": {"id": "09014c72-f5d4-4f50-b1c8-b4ca85795de9", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-696815363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "298bd92edec8447a9377fac94e783f16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d722a23-d4", "ovs_interfaceid": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.805169] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.808025] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.736s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.808365] env[61947]: DEBUG nova.objects.instance [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lazy-loading 'resources' on Instance uuid d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.863619] env[61947]: INFO nova.scheduler.client.report [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Deleted allocations for instance 74e03575-297e-4e08-9236-98d8be80b546 [ 859.956904] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6aa384ea-9223-456a-b090-592c3697c122 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.960s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.049220] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224462, 'name': CreateVM_Task, 'duration_secs': 1.539027} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.049220] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 860.049220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.049220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.049220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.049220] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f51b8ddf-6e96-491e-8c71-05ff21f08bb8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.056183] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 860.056183] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522110c6-25d8-8075-4f27-f738d950b75c" [ 860.056183] env[61947]: _type = "Task" [ 860.056183] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.071829] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522110c6-25d8-8075-4f27-f738d950b75c, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.073013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.075653] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.075653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.075653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.075653] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.075653] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9bc98b2-a1b6-48e1-8566-a70a485719a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.081882] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224463, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.529553} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.082864] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 8c342d64-75f6-4b7d-829d-889eb3b6122f/8c342d64-75f6-4b7d-829d-889eb3b6122f.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 860.083056] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.083256] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0d10f50-0aae-4874-a09d-203bc70b8e3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.090725] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.090935] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.092766] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ed47fe4-2aea-4552-9c86-8fd18d19231d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.097308] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 860.097308] env[61947]: value = "task-1224464" [ 860.097308] env[61947]: _type = "Task" [ 860.097308] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.104785] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 860.104785] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52aa08f7-adff-02ea-2484-849120ee5197" [ 860.104785] env[61947]: _type = "Task" [ 860.104785] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.114635] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.122407] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52aa08f7-adff-02ea-2484-849120ee5197, 'name': SearchDatastore_Task, 'duration_secs': 0.01307} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.122816] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0db65396-fa24-4db6-a5fc-08471e99f2fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.129981] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Received event network-vif-plugged-7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.130249] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Acquiring lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.130473] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.130742] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.130806] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] No waiting events found dispatching network-vif-plugged-7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.130969] env[61947]: WARNING nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Received unexpected event network-vif-plugged-7d722a23-d40a-41cc-8e2a-c4456d5b456c for instance with vm_state building and task_state spawning. [ 860.131151] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Received event network-changed-7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.131313] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Refreshing instance network info cache due to event network-changed-7d722a23-d40a-41cc-8e2a-c4456d5b456c. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 860.131817] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Acquiring lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.137180] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 860.137180] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c7be26-99da-b798-96cf-98493576a82d" [ 860.137180] env[61947]: _type = "Task" [ 860.137180] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.149141] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c7be26-99da-b798-96cf-98493576a82d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.176065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Releasing lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.176525] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Instance network_info: |[{"id": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "address": "fa:16:3e:aa:64:bf", "network": {"id": "09014c72-f5d4-4f50-b1c8-b4ca85795de9", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-696815363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "298bd92edec8447a9377fac94e783f16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d722a23-d4", "ovs_interfaceid": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.176948] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Acquired lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.178030] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Refreshing network info cache for port 7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.179902] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:64:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d722a23-d40a-41cc-8e2a-c4456d5b456c', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.188902] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Creating folder: Project (298bd92edec8447a9377fac94e783f16). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.190255] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1014d4a-7dff-4eeb-ab36-9931b7145ec2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.204014] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Created folder: Project (298bd92edec8447a9377fac94e783f16) in parent group-v264556. [ 860.207711] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Creating folder: Instances. Parent ref: group-v264652. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.207711] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8c0913a-a1b3-4e9c-837f-163b5088f138 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.217963] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Created folder: Instances in parent group-v264652. [ 860.218095] env[61947]: DEBUG oslo.service.loopingcall [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.218251] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.218556] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97ab9aee-ba29-44e8-9009-d23b396b979c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.245674] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.245674] env[61947]: value = "task-1224467" [ 860.245674] env[61947]: _type = "Task" [ 860.245674] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.256153] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224467, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.380265] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f4ba4ace-0b35-420e-9b30-b5f0523dbcae tempest-ServersNegativeTestMultiTenantJSON-39275672 tempest-ServersNegativeTestMultiTenantJSON-39275672-project-member] Lock "74e03575-297e-4e08-9236-98d8be80b546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.889s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.608660] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088087} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.608886] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.609695] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5187978b-adf6-422c-9dc7-3764f379dac6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.638755] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 8c342d64-75f6-4b7d-829d-889eb3b6122f/8c342d64-75f6-4b7d-829d-889eb3b6122f.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.642037] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd211f75-2c1c-46b0-a7c9-9438dda7d83d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.670980] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c7be26-99da-b798-96cf-98493576a82d, 'name': SearchDatastore_Task, 'duration_secs': 0.013394} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.670980] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.670980] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4f0b9636-4793-4b1c-8b2b-7f31a15669bb/4f0b9636-4793-4b1c-8b2b-7f31a15669bb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.670980] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 860.670980] env[61947]: value = "task-1224468" [ 860.670980] env[61947]: _type = "Task" [ 860.670980] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.673561] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f48b4ee4-344c-4c93-a6eb-bf50862e6b62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.689645] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224468, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.694108] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 860.694108] env[61947]: value = "task-1224469" [ 860.694108] env[61947]: _type = "Task" [ 860.694108] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.695653] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f146d9a-881d-480c-97d8-973f48c0c15e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.704831] env[61947]: DEBUG nova.compute.manager [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.704831] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecb94d5-b5bf-4eed-be96-d53a5963afa4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.714424] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2764f9a0-c616-46e6-b989-0dcd0746e763 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.718130] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.760228] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a8aff5-f82f-4861-8c8c-2e8a324f45d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.774307] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224467, 'name': CreateVM_Task, 'duration_secs': 0.466317} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.777172] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0610647-847c-4f56-8fca-846ea56501fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.781432] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 860.782744] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.782852] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.783180] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.783873] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c49c3b9-c707-4c9b-9ce2-7cb51dc4ec1c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.795758] env[61947]: DEBUG nova.compute.provider_tree [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.799188] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 860.799188] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f5d2c4-0f96-0f7b-5d5c-66466ce21156" [ 860.799188] env[61947]: _type = "Task" [ 860.799188] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.809394] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f5d2c4-0f96-0f7b-5d5c-66466ce21156, 'name': SearchDatastore_Task, 'duration_secs': 0.011973} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.809394] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.809931] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.809931] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.809931] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.810121] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.810484] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-777339e7-dc41-4fa6-a9b7-96a8a4fc3e1e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.825030] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.825030] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 860.825030] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91deba6e-a36b-47c6-b437-3be7ca60d263 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.835405] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 860.835405] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52579671-72d0-844c-6958-b32ee60037a8" [ 860.835405] env[61947]: _type = "Task" [ 860.835405] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.843576] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52579671-72d0-844c-6958-b32ee60037a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.190580] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224468, 'name': ReconfigVM_Task, 'duration_secs': 0.454509} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.191129] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 8c342d64-75f6-4b7d-829d-889eb3b6122f/8c342d64-75f6-4b7d-829d-889eb3b6122f.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.191889] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-293aa8cb-9f5b-4916-ae51-03977001d45b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.204979] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 861.204979] env[61947]: value = "task-1224470" [ 861.204979] env[61947]: _type = "Task" [ 861.204979] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.214911] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500217} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.215741] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4f0b9636-4793-4b1c-8b2b-7f31a15669bb/4f0b9636-4793-4b1c-8b2b-7f31a15669bb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.216116] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.216465] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-078cc091-361a-4b1c-a327-e492bd4d25f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.222169] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224470, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.223733] env[61947]: INFO nova.compute.manager [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] instance snapshotting [ 861.231256] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 861.231256] env[61947]: value = "task-1224471" [ 861.231256] env[61947]: _type = "Task" [ 861.231256] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.232142] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736f8f42-189c-4f18-9dee-f1f648a270ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.248113] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.270209] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aabcab8-8144-42cf-9fb2-c4dbf9216bec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.302342] env[61947]: DEBUG nova.scheduler.client.report [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.344255] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52579671-72d0-844c-6958-b32ee60037a8, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.345140] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42fb4058-0f80-4065-86aa-594c07fc87f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.352851] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 861.352851] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52060e37-4b2d-2ab0-6cce-626b2aba111f" [ 861.352851] env[61947]: _type = "Task" [ 861.352851] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.365144] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52060e37-4b2d-2ab0-6cce-626b2aba111f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.416513] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Updated VIF entry in instance network info cache for port 7d722a23-d40a-41cc-8e2a-c4456d5b456c. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.416961] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Updating instance_info_cache with network_info: [{"id": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "address": "fa:16:3e:aa:64:bf", "network": {"id": "09014c72-f5d4-4f50-b1c8-b4ca85795de9", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-696815363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "298bd92edec8447a9377fac94e783f16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d722a23-d4", "ovs_interfaceid": "7d722a23-d40a-41cc-8e2a-c4456d5b456c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.716283] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224470, 'name': Rename_Task, 'duration_secs': 0.196101} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.716889] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.717198] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f46d40fa-179b-40ce-9dfb-a34bdcb68d48 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.725799] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 861.725799] env[61947]: value = "task-1224472" [ 861.725799] env[61947]: _type = "Task" [ 861.725799] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.735299] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.745156] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068556} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.745465] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.746298] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb938f6-9127-4f33-bbef-783c832fb62d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.773763] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 4f0b9636-4793-4b1c-8b2b-7f31a15669bb/4f0b9636-4793-4b1c-8b2b-7f31a15669bb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.774469] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c739421d-6e7e-4805-a460-817879165168 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.790113] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 861.790493] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9b21e6fb-f060-42e4-91e9-1105b4804102 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.800054] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 861.800054] env[61947]: value = "task-1224473" [ 861.800054] env[61947]: _type = "Task" [ 861.800054] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.801548] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 861.801548] env[61947]: value = "task-1224474" [ 861.801548] env[61947]: _type = "Task" [ 861.801548] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.805529] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.998s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.811447] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.117s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.811826] env[61947]: DEBUG nova.objects.instance [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'resources' on Instance uuid 7efd65d9-67f9-497d-9e42-51d8973aa695 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.820415] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224473, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.825031] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224474, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.839891] env[61947]: INFO nova.scheduler.client.report [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleted allocations for instance d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd [ 861.865528] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52060e37-4b2d-2ab0-6cce-626b2aba111f, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.865810] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.866192] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ee3b9509-5437-41b3-b612-91d148a338b6/ee3b9509-5437-41b3-b612-91d148a338b6.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 861.867157] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62427415-795c-48fd-a831-2edb05121133 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.877269] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 861.877269] env[61947]: value = "task-1224475" [ 861.877269] env[61947]: _type = "Task" [ 861.877269] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.891548] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.920312] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Releasing lock "refresh_cache-ee3b9509-5437-41b3-b612-91d148a338b6" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.922102] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Received event network-changed-4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 861.922102] env[61947]: DEBUG nova.compute.manager [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Refreshing instance network info cache due to event network-changed-4024dcff-f052-4a2e-9fab-c3d722069bbe. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 861.922102] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Acquiring lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.922102] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Acquired lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.922102] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Refreshing network info cache for port 4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.179775] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 862.180826] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2fcf33-7c4a-4488-98ae-1b4c78365a82 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.193885] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 862.193885] env[61947]: ERROR oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk due to incomplete transfer. [ 862.193885] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4b99cec0-7523-472e-a3f6-581a7fc915bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.205027] env[61947]: DEBUG oslo_vmware.rw_handles [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234fe22-725e-baa6-a3c9-19e3f7ea7786/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 862.205027] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploaded image 1cb3a519-7891-4eeb-9474-821506e882fb to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 862.207313] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 862.207884] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-504003dc-dcd7-45f2-a136-47ede232cdca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.219783] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 862.219783] env[61947]: value = "task-1224476" [ 862.219783] env[61947]: _type = "Task" [ 862.219783] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.234735] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224476, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.243014] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224472, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.332418] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224473, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.336596] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224474, 'name': ReconfigVM_Task, 'duration_secs': 0.506469} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.336894] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 4f0b9636-4793-4b1c-8b2b-7f31a15669bb/4f0b9636-4793-4b1c-8b2b-7f31a15669bb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.338453] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10a4369a-9312-4786-8e6c-f4c8d669b1bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.348652] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 862.348652] env[61947]: value = "task-1224477" [ 862.348652] env[61947]: _type = "Task" [ 862.348652] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.349926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-54b8a1c1-1dc2-41d7-9230-fe327052205e tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.943s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.367899] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224477, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.399377] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224475, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492608} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.400114] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ee3b9509-5437-41b3-b612-91d148a338b6/ee3b9509-5437-41b3-b612-91d148a338b6.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 862.400114] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.401177] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d3523dc-7ef0-4b16-b8bb-278a0d4e029b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.409710] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 862.409710] env[61947]: value = "task-1224478" [ 862.409710] env[61947]: _type = "Task" [ 862.409710] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.431516] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224478, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.682353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b07d53a-ac76-4686-9ce9-8c0ba824e122 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.693619] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1b6bf1-85f7-4141-ad53-0386686c516b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.742051] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72aab160-7157-4b55-bccd-8b55597eba36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.754195] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224476, 'name': Destroy_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.759383] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba2e439-80e3-497b-8037-4c7c9e9ad549 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.763712] env[61947]: DEBUG oslo_vmware.api [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224472, 'name': PowerOnVM_Task, 'duration_secs': 0.702524} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.763995] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 862.764249] env[61947]: INFO nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Took 10.09 seconds to spawn the instance on the hypervisor. [ 862.764456] env[61947]: DEBUG nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.765603] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d382a8-a193-49ed-ab4e-5e24013b5f9e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.775806] env[61947]: DEBUG nova.compute.provider_tree [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.812943] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224473, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.864227] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224477, 'name': Rename_Task, 'duration_secs': 0.267178} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.864525] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.868029] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f4eb701-39d7-43fa-b96e-660862f915f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.872961] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 862.872961] env[61947]: value = "task-1224479" [ 862.872961] env[61947]: _type = "Task" [ 862.872961] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.882153] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.917085] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updated VIF entry in instance network info cache for port 4024dcff-f052-4a2e-9fab-c3d722069bbe. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.917085] env[61947]: DEBUG nova.network.neutron [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updating instance_info_cache with network_info: [{"id": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "address": "fa:16:3e:91:65:29", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4024dcff-f0", "ovs_interfaceid": "4024dcff-f052-4a2e-9fab-c3d722069bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.921836] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224478, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079842} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.922350] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.923155] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2202b153-4ae1-4d98-a655-1b43f60e95b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.949657] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] ee3b9509-5437-41b3-b612-91d148a338b6/ee3b9509-5437-41b3-b612-91d148a338b6.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.950863] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d52f7ad-2949-4d07-b9b9-d1a941b638e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.971859] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 862.971859] env[61947]: value = "task-1224480" [ 862.971859] env[61947]: _type = "Task" [ 862.971859] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.982241] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224480, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.243283] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224476, 'name': Destroy_Task, 'duration_secs': 0.781826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.243283] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroyed the VM [ 863.243526] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 863.243798] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a839a5bc-0f2b-48ba-b72f-fe6304243157 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.252290] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 863.252290] env[61947]: value = "task-1224481" [ 863.252290] env[61947]: _type = "Task" [ 863.252290] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.261331] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.279652] env[61947]: DEBUG nova.scheduler.client.report [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.293476] env[61947]: INFO nova.compute.manager [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Took 29.65 seconds to build instance. [ 863.314549] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224473, 'name': CreateSnapshot_Task, 'duration_secs': 1.346634} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.314835] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 863.315622] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190fa46e-972d-465b-8418-8a692fe38cb8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.384640] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.423457] env[61947]: DEBUG oslo_concurrency.lockutils [req-5b523abe-a1d2-4bd7-9a0d-ebbea429f028 req-84b3aead-f32e-4f44-951d-88a9143e5f7d service nova] Releasing lock "refresh_cache-bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.484403] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224480, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.764461] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.785526] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.788685] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.101s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.788685] env[61947]: DEBUG nova.objects.instance [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lazy-loading 'resources' on Instance uuid 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.795749] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3b007179-d4cc-4e4f-8beb-01af87a8c84a tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.739s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.816400] env[61947]: INFO nova.scheduler.client.report [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance 7efd65d9-67f9-497d-9e42-51d8973aa695 [ 863.836372] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 863.836672] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e8abe640-aa3e-4ae1-ab70-07c4200a2dc6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.847656] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 863.847656] env[61947]: value = "task-1224482" [ 863.847656] env[61947]: _type = "Task" [ 863.847656] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.857823] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224482, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.889992] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.915244] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.915531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.915786] env[61947]: INFO nova.compute.manager [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Shelving [ 863.994223] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224480, 'name': ReconfigVM_Task, 'duration_secs': 0.935052} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.994540] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Reconfigured VM instance instance-00000044 to attach disk [datastore2] ee3b9509-5437-41b3-b612-91d148a338b6/ee3b9509-5437-41b3-b612-91d148a338b6.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.995203] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c063393-f01b-4b79-8e58-050a959e8790 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.004477] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 864.004477] env[61947]: value = "task-1224483" [ 864.004477] env[61947]: _type = "Task" [ 864.004477] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.014500] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.264841] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.327837] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e8db6cfc-22d0-41d1-9d70-ec617134e452 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "7efd65d9-67f9-497d-9e42-51d8973aa695" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.992s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.364703] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224482, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.386025] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.427430] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.427710] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd093929-3f29-43cc-b8d5-46521c5d6315 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.438692] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 864.438692] env[61947]: value = "task-1224484" [ 864.438692] env[61947]: _type = "Task" [ 864.438692] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.454419] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224484, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.518949] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.658497] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4701c2c7-bb6e-4349-9bcc-9873541781f7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.666819] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fbefa8-1493-4bef-9734-4a8ee2a3ad86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.699781] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acac08b-6415-4c51-ba31-e77ce8b4cc49 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.708370] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b7d9d0-4502-42a3-b95b-b84751bf6318 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.723250] env[61947]: DEBUG nova.compute.provider_tree [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.764425] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.861489] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224482, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.887154] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.952837] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224484, 'name': PowerOffVM_Task, 'duration_secs': 0.180612} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.952837] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.953774] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc8e086-070e-436b-a5ee-fe81a31870d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.977261] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f93bcb-293a-4886-b743-80277bb2fffe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.020657] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.179021] env[61947]: DEBUG nova.compute.manager [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Received event network-changed-1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.179021] env[61947]: DEBUG nova.compute.manager [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Refreshing instance network info cache due to event network-changed-1a026df8-fa69-439e-805b-5dad5c772246. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 865.179021] env[61947]: DEBUG oslo_concurrency.lockutils [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] Acquiring lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.179021] env[61947]: DEBUG oslo_concurrency.lockutils [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] Acquired lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.179021] env[61947]: DEBUG nova.network.neutron [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Refreshing network info cache for port 1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.227621] env[61947]: DEBUG nova.scheduler.client.report [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.265507] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.364613] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224482, 'name': CloneVM_Task} progress is 95%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.391214] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.492990] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 865.492990] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e8b349fd-7304-42d7-b099-f64322a2e160 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.502380] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 865.502380] env[61947]: value = "task-1224485" [ 865.502380] env[61947]: _type = "Task" [ 865.502380] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.515995] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224485, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.519656] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.735887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.738514] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.473s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.740395] env[61947]: INFO nova.compute.claims [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.765848] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.768798] env[61947]: INFO nova.scheduler.client.report [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Deleted allocations for instance 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6 [ 865.851654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.851654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.865955] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224482, 'name': CloneVM_Task, 'duration_secs': 1.844835} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.869024] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Created linked-clone VM from snapshot [ 865.869814] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32319b7e-1229-4553-8ef9-88e6812dd70c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.883120] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Uploading image 13296250-6feb-41b0-b80f-26c6dd557dde {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 865.898890] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.925954] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 865.925954] env[61947]: value = "vm-264656" [ 865.925954] env[61947]: _type = "VirtualMachine" [ 865.925954] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 865.926866] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-de9776bd-64b1-4154-8751-ac2a1131cb82 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.936775] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease: (returnval){ [ 865.936775] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520e1502-5100-3396-b5a8-9937c28f5bf4" [ 865.936775] env[61947]: _type = "HttpNfcLease" [ 865.936775] env[61947]: } obtained for exporting VM: (result){ [ 865.936775] env[61947]: value = "vm-264656" [ 865.936775] env[61947]: _type = "VirtualMachine" [ 865.936775] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 865.936999] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the lease: (returnval){ [ 865.936999] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520e1502-5100-3396-b5a8-9937c28f5bf4" [ 865.936999] env[61947]: _type = "HttpNfcLease" [ 865.936999] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 865.947994] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 865.947994] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520e1502-5100-3396-b5a8-9937c28f5bf4" [ 865.947994] env[61947]: _type = "HttpNfcLease" [ 865.947994] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 866.018731] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224485, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.022977] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.210722] env[61947]: DEBUG nova.network.neutron [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updated VIF entry in instance network info cache for port 1a026df8-fa69-439e-805b-5dad5c772246. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.211143] env[61947]: DEBUG nova.network.neutron [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updating instance_info_cache with network_info: [{"id": "1a026df8-fa69-439e-805b-5dad5c772246", "address": "fa:16:3e:5d:8b:d0", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a026df8-fa", "ovs_interfaceid": "1a026df8-fa69-439e-805b-5dad5c772246", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.273017] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.284055] env[61947]: DEBUG oslo_concurrency.lockutils [None req-691b8460-4cf4-46c6-9d7b-be4b92b6f47f tempest-ServersTestManualDisk-1046515137 tempest-ServersTestManualDisk-1046515137-project-member] Lock "7612f830-fed2-4dd7-ba8f-f792e1aa5ac6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.277s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.359057] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.399804] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.448878] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.448878] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520e1502-5100-3396-b5a8-9937c28f5bf4" [ 866.448878] env[61947]: _type = "HttpNfcLease" [ 866.448878] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 866.449350] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 866.449350] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520e1502-5100-3396-b5a8-9937c28f5bf4" [ 866.449350] env[61947]: _type = "HttpNfcLease" [ 866.449350] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 866.450454] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e4390d-cc6a-45b4-9f0a-77ffbbed0071 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.458035] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 866.458035] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 866.542700] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.546066] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224485, 'name': CreateSnapshot_Task, 'duration_secs': 0.814023} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.546371] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 866.547141] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a346cfbc-b9c9-453f-b651-805a0fabf7ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.585132] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8838a7aa-803a-4a4d-9e5d-8de48a42d79b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.714067] env[61947]: DEBUG oslo_concurrency.lockutils [req-21b7ee29-f849-4097-98ff-14e506a6e7b2 req-69bbaf30-1545-4ad2-a5f0-fac6d271742d service nova] Releasing lock "refresh_cache-8c342d64-75f6-4b7d-829d-889eb3b6122f" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.766462] env[61947]: DEBUG oslo_vmware.api [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224481, 'name': RemoveSnapshot_Task, 'duration_secs': 3.495235} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.766755] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 866.766992] env[61947]: INFO nova.compute.manager [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 16.61 seconds to snapshot the instance on the hypervisor. [ 866.897554] env[61947]: DEBUG oslo_vmware.api [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224479, 'name': PowerOnVM_Task, 'duration_secs': 3.818161} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.897554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.898706] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 866.898947] env[61947]: INFO nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Took 11.69 seconds to spawn the instance on the hypervisor. [ 866.899170] env[61947]: DEBUG nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.900475] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472072fd-9cad-4fea-94ad-c2f89bc89364 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.043221] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224483, 'name': Rename_Task, 'duration_secs': 2.599217} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.044324] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.045035] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4da4d2b-a6d4-4490-9822-3e620ac54e4e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.054355] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 867.054355] env[61947]: value = "task-1224487" [ 867.054355] env[61947]: _type = "Task" [ 867.054355] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.068024] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 867.072078] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-67feb580-b5fd-481b-9abf-717fcc383cf7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.090850] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.091346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.093527] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224487, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.097349] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 867.097349] env[61947]: value = "task-1224488" [ 867.097349] env[61947]: _type = "Task" [ 867.097349] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.117841] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224488, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.150125] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2835a8e4-7416-4dc2-b2ab-7ad7b658842d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.162069] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79f302c-c00b-4cfa-bf81-7f59cd811948 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.197203] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82327a4b-0ce2-4464-857e-08c3f41b69cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.206181] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86653423-fbac-40dc-a43f-23f81239ed7e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.222674] env[61947]: DEBUG nova.compute.provider_tree [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.318024] env[61947]: DEBUG nova.compute.manager [None req-2df38622-6979-4129-86ae-15e8c3545fbb tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Found 1 images (rotation: 2) {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 867.421852] env[61947]: INFO nova.compute.manager [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Took 32.44 seconds to build instance. [ 867.569909] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224487, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.595153] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.610302] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224488, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.729197] env[61947]: DEBUG nova.scheduler.client.report [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.926671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2d2dbb25-e6e6-4233-b450-939dd36d919f tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.360s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.070698] env[61947]: DEBUG oslo_vmware.api [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224487, 'name': PowerOnVM_Task, 'duration_secs': 0.622352} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.071369] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.074885] env[61947]: INFO nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Took 10.32 seconds to spawn the instance on the hypervisor. [ 868.074885] env[61947]: DEBUG nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.074885] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc202186-0ce0-4fe6-a7bc-ae7c59d015b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.119896] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224488, 'name': CloneVM_Task} progress is 95%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.129788] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.238437] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.239177] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.249081] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 24.824s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.602685] env[61947]: INFO nova.compute.manager [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Took 31.34 seconds to build instance. [ 868.620326] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224488, 'name': CloneVM_Task, 'duration_secs': 1.374696} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.620885] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Created linked-clone VM from snapshot [ 868.622383] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9fccc4-cc07-4e43-918d-23cd1e87f163 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.630454] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Uploading image ef8d583a-38fd-4d74-9ed7-f1718e80cace {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 868.659972] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 868.659972] env[61947]: value = "vm-264658" [ 868.659972] env[61947]: _type = "VirtualMachine" [ 868.659972] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 868.659972] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cdf725ba-51a5-4f09-9893-b6eef6fe6c47 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.667022] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lease: (returnval){ [ 868.667022] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528189f8-077a-2f3b-d8b5-eaf8b66b5d7a" [ 868.667022] env[61947]: _type = "HttpNfcLease" [ 868.667022] env[61947]: } obtained for exporting VM: (result){ [ 868.667022] env[61947]: value = "vm-264658" [ 868.667022] env[61947]: _type = "VirtualMachine" [ 868.667022] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 868.667022] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the lease: (returnval){ [ 868.667022] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528189f8-077a-2f3b-d8b5-eaf8b66b5d7a" [ 868.667022] env[61947]: _type = "HttpNfcLease" [ 868.667022] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 868.674376] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.674376] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528189f8-077a-2f3b-d8b5-eaf8b66b5d7a" [ 868.674376] env[61947]: _type = "HttpNfcLease" [ 868.674376] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 868.751863] env[61947]: DEBUG nova.compute.utils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.761034] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Not allocating networking since 'none' was specified. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 869.017346] env[61947]: DEBUG nova.compute.manager [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.018521] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920e4c37-390c-4e9c-9d90-1860b73c89c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.086015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.086015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.086015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.086015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.086015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.088073] env[61947]: INFO nova.compute.manager [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Terminating instance [ 869.090060] env[61947]: DEBUG nova.compute.manager [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 869.090306] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.091234] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d08cfe-c8c4-4b52-bc23-aa5e05a7e972 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.101292] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.102223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-57495482-0725-4588-8411-a9f7e9f50c75 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.476s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.102440] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c100737-a63c-440f-b4e8-c8d8ef53798f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.110159] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 869.110159] env[61947]: value = "task-1224490" [ 869.110159] env[61947]: _type = "Task" [ 869.110159] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.119713] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224490, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.175610] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.175610] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528189f8-077a-2f3b-d8b5-eaf8b66b5d7a" [ 869.175610] env[61947]: _type = "HttpNfcLease" [ 869.175610] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 869.175926] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 869.175926] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528189f8-077a-2f3b-d8b5-eaf8b66b5d7a" [ 869.175926] env[61947]: _type = "HttpNfcLease" [ 869.175926] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 869.176690] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e26f77-968d-4019-9a0a-4d0e8cd0a3c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.184647] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 869.184841] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 869.262075] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.266987] env[61947]: INFO nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating resource usage from migration 89e41994-bff5-4166-a6a9-10258b48d1d3 [ 869.267186] env[61947]: INFO nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating resource usage from migration 91d49010-544b-497e-ad54-e402ab5bdd95 [ 869.278281] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6a762623-cf88-4d75-be00-5be036fe4a01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.293030] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ab027a9d-f51d-4cb6-8aaf-ebd507548657 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.293312] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.293488] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f68a8537-5bae-4b50-b0ff-fa9b4a89bd40 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 869.293638] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f64b9f7f-89a9-4fb2-9c76-13cc591b9d51 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: WARNING nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7d7f267a-3418-4b9b-9589-c915d5194f0d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9bc18645-c837-4dab-8e05-fbcf0d40812f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8c342d64-75f6-4b7d-829d-889eb3b6122f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4f0b9636-4793-4b1c-8b2b-7f31a15669bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ee3b9509-5437-41b3-b612-91d148a338b6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance e0eebf59-c84e-4462-b280-d783a04525e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Migration 91d49010-544b-497e-ad54-e402ab5bdd95 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 869.294821] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance eede4f34-5ac4-475c-a74b-a98327f648fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.529627] env[61947]: INFO nova.compute.manager [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] instance snapshotting [ 869.530894] env[61947]: DEBUG nova.objects.instance [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.603263] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "interface-ee3b9509-5437-41b3-b612-91d148a338b6-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.603470] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "interface-ee3b9509-5437-41b3-b612-91d148a338b6-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.603915] env[61947]: DEBUG nova.objects.instance [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lazy-loading 'flavor' on Instance uuid ee3b9509-5437-41b3-b612-91d148a338b6 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.624052] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224490, 'name': PowerOffVM_Task, 'duration_secs': 0.282212} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.624052] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.624258] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.624473] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6310633d-30db-4ae8-9b97-fda93e410402 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.693374] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.693561] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.693722] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Deleting the datastore file [datastore2] 4f0b9636-4793-4b1c-8b2b-7f31a15669bb {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.694388] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-461d7289-0e63-409d-a95f-8766a99710a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.704953] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for the task: (returnval){ [ 869.704953] env[61947]: value = "task-1224492" [ 869.704953] env[61947]: _type = "Task" [ 869.704953] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.714654] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.802933] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 7f308223-6c3c-4c62-8401-d7a44eeb69d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 869.803139] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Migration 89e41994-bff5-4166-a6a9-10258b48d1d3 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 869.803299] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 72818e40-624b-4c04-888b-bb622f7f96d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.038959] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e53c3d-28f6-44e0-bf3f-b3e7c22fe1b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.059179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1c0ea5-6e99-4593-af40-42c2ea9a33fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.109093] env[61947]: DEBUG nova.objects.instance [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lazy-loading 'pci_requests' on Instance uuid ee3b9509-5437-41b3-b612-91d148a338b6 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.215627] env[61947]: DEBUG oslo_vmware.api [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Task: {'id': task-1224492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266692} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.215938] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.216163] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.216476] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.216835] env[61947]: INFO nova.compute.manager [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 870.217288] env[61947]: DEBUG oslo.service.loopingcall [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.217596] env[61947]: DEBUG nova.compute.manager [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.217748] env[61947]: DEBUG nova.network.neutron [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.278187] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.306621] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4f313bb5-5df4-4fb0-9736-95f4a519651d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 870.571375] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 870.571957] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4dcc0632-56b3-404a-bdce-02cb36fad9a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.582760] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 870.582760] env[61947]: value = "task-1224493" [ 870.582760] env[61947]: _type = "Task" [ 870.582760] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.595384] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224493, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.613922] env[61947]: DEBUG nova.objects.base [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 870.613922] env[61947]: DEBUG nova.network.neutron [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.707459] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cbc29efe-ab40-47f4-8fbc-a7fcbcc8f2bd tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "interface-ee3b9509-5437-41b3-b612-91d148a338b6-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.102s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.813860] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9cfa7c52-2250-49ce-a1c8-baad6644f824 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 870.813860] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 870.813860] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 870.998464] env[61947]: DEBUG nova.network.neutron [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.104040] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224493, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.182292] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6293081-580f-47fd-a2f1-87bde2c356b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.189704] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23fbcdfa-6ebc-4d00-9999-0d2df7dcae06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.234968] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5401c4e4-6a91-4eeb-add2-2ed9b8febe81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.244585] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d37a76-8eb7-45c3-9457-eb62f0023c4a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.259744] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.502024] env[61947]: INFO nova.compute.manager [-] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Took 1.28 seconds to deallocate network for instance. [ 871.596127] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224493, 'name': CreateSnapshot_Task, 'duration_secs': 0.77718} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.596370] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 871.597163] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0be315-e8e8-40e0-87b0-ba692f4fa621 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.763308] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.010433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.117803] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 872.118295] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-746d27a6-21a9-4921-be74-d9a1eb9c0bf5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.127851] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 872.127851] env[61947]: value = "task-1224494" [ 872.127851] env[61947]: _type = "Task" [ 872.127851] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.139095] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224494, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.268498] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 872.268809] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.023s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.269202] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.720s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.269503] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.272138] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.391s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.272431] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.274691] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 27.270s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.305642] env[61947]: INFO nova.scheduler.client.report [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted allocations for instance f68a8537-5bae-4b50-b0ff-fa9b4a89bd40 [ 872.307329] env[61947]: INFO nova.scheduler.client.report [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Deleted allocations for instance f64b9f7f-89a9-4fb2-9c76-13cc591b9d51 [ 872.643784] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224494, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.650582] env[61947]: DEBUG nova.compute.manager [req-20c2ab67-802e-42c2-9ca5-64d8829dd028 req-152ffbeb-e567-4e87-9fc0-6dc62c6cb9e6 service nova] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Received event network-vif-deleted-d47b903b-da5e-4363-9b20-4d4238cdc3e2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.781535] env[61947]: INFO nova.compute.claims [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.822720] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fed565c-dd6c-4ca8-82d1-46a3dc629f91 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f68a8537-5bae-4b50-b0ff-fa9b4a89bd40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.513s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.836637] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a208b74e-e779-49c9-9249-372637fef715 tempest-MultipleCreateTestJSON-480841736 tempest-MultipleCreateTestJSON-480841736-project-member] Lock "f64b9f7f-89a9-4fb2-9c76-13cc591b9d51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.434s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.142361] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224494, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.291021] env[61947]: INFO nova.compute.resource_tracker [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating resource usage from migration 91d49010-544b-497e-ad54-e402ab5bdd95 [ 873.595197] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00aa1d8c-76c8-44ee-b373-05361edfff4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.610161] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1dc2d3-08f9-4187-b9ed-637c08841c78 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.651126] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "ee3b9509-5437-41b3-b612-91d148a338b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.651438] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.651767] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.651889] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.652086] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.658558] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d02037e-8671-4598-b340-17b4a2401ff8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.661672] env[61947]: INFO nova.compute.manager [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Terminating instance [ 873.665857] env[61947]: DEBUG nova.compute.manager [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.665857] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 873.667657] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae256f0b-9ec0-4878-98e4-5624614cb23d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.675984] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224494, 'name': CloneVM_Task, 'duration_secs': 1.490298} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.677224] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created linked-clone VM from snapshot [ 873.678667] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd0e081-04c5-471d-bd80-c4912440072b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.689325] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fed31c-0bd4-424d-820e-9238318016df {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.693713] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.694611] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96a1ce69-b6b7-4fa7-bdf3-fe6e7fa0a56d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.710904] env[61947]: DEBUG nova.compute.provider_tree [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.718511] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploading image 2cd5eb90-64f8-4008-92f1-e73346f0afac {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 873.721929] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 873.721929] env[61947]: value = "task-1224495" [ 873.721929] env[61947]: _type = "Task" [ 873.721929] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.732095] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.223293] env[61947]: DEBUG nova.scheduler.client.report [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.247613] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224495, 'name': PowerOffVM_Task, 'duration_secs': 0.286689} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.247613] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 874.247613] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 874.248156] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e186eb5f-1f1d-4348-a3bc-8b9757b32573 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.314107] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 874.314107] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 874.314107] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Deleting the datastore file [datastore2] ee3b9509-5437-41b3-b612-91d148a338b6 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.317266] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d5386a5-1952-4398-8139-7a0372da3526 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.324020] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for the task: (returnval){ [ 874.324020] env[61947]: value = "task-1224497" [ 874.324020] env[61947]: _type = "Task" [ 874.324020] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.331716] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.739707] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.465s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.740346] env[61947]: INFO nova.compute.manager [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Migrating [ 874.740346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.740346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.741768] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.929s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.741996] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.744232] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.885s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.745869] env[61947]: INFO nova.compute.claims [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.749406] env[61947]: INFO nova.compute.rpcapi [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 874.750035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.779471] env[61947]: INFO nova.scheduler.client.report [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Deleted allocations for instance 7d7f267a-3418-4b9b-9589-c915d5194f0d [ 874.842566] env[61947]: DEBUG oslo_vmware.api [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Task: {'id': task-1224497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.42224} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.842933] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.843367] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 874.843704] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 874.843986] env[61947]: INFO nova.compute.manager [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 874.844369] env[61947]: DEBUG oslo.service.loopingcall [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.845463] env[61947]: DEBUG nova.compute.manager [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.845601] env[61947]: DEBUG nova.network.neutron [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.266414] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.266414] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.266793] env[61947]: DEBUG nova.network.neutron [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.289013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-798379da-bedb-4c66-8133-b8c86b5c3682 tempest-ServersAaction247Test-1035387076 tempest-ServersAaction247Test-1035387076-project-member] Lock "7d7f267a-3418-4b9b-9589-c915d5194f0d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.259s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.303307] env[61947]: DEBUG nova.compute.manager [req-394d113f-b355-4794-bbb6-30081405dfc5 req-87f0694e-594e-4778-8d28-9fafbc3bc878 service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Received event network-vif-deleted-7d722a23-d40a-41cc-8e2a-c4456d5b456c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.304048] env[61947]: INFO nova.compute.manager [req-394d113f-b355-4794-bbb6-30081405dfc5 req-87f0694e-594e-4778-8d28-9fafbc3bc878 service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Neutron deleted interface 7d722a23-d40a-41cc-8e2a-c4456d5b456c; detaching it from the instance and deleting it from the info cache [ 875.304256] env[61947]: DEBUG nova.network.neutron [req-394d113f-b355-4794-bbb6-30081405dfc5 req-87f0694e-594e-4778-8d28-9fafbc3bc878 service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.344521] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.344722] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.344876] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.345067] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.345233] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.345358] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.345577] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.345732] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.345943] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.346063] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.346236] env[61947]: DEBUG nova.virt.hardware [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.349649] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685de36b-892f-4bf6-bde7-90ff15cf6120 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.360356] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 875.360356] env[61947]: value = "vm-264660" [ 875.360356] env[61947]: _type = "VirtualMachine" [ 875.360356] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 875.362044] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14847e3-3e16-495c-93e8-30f650352286 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.369102] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-452cfb1f-b374-44ca-83cb-2236576576ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.386410] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.395153] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Creating folder: Project (c75d2366d4e1483b99a9c4eb9bdac9a1). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 875.395649] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 875.395927] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8abd7d4-0e91-45a6-aa54-1a437ee2847b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.397811] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease: (returnval){ [ 875.397811] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e1393f-eb49-3684-48f8-2093e52c3b1c" [ 875.397811] env[61947]: _type = "HttpNfcLease" [ 875.397811] env[61947]: } obtained for exporting VM: (result){ [ 875.397811] env[61947]: value = "vm-264660" [ 875.397811] env[61947]: _type = "VirtualMachine" [ 875.397811] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 875.398143] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the lease: (returnval){ [ 875.398143] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e1393f-eb49-3684-48f8-2093e52c3b1c" [ 875.398143] env[61947]: _type = "HttpNfcLease" [ 875.398143] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 875.398901] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684ca087-c502-469a-aac7-e55ee2272073 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.408400] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 875.409037] env[61947]: ERROR oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk due to incomplete transfer. [ 875.411753] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f9a53b52-ad35-40bc-9994-87374ae1e4f8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.415038] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.415038] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e1393f-eb49-3684-48f8-2093e52c3b1c" [ 875.415038] env[61947]: _type = "HttpNfcLease" [ 875.415038] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 875.415038] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Created folder: Project (c75d2366d4e1483b99a9c4eb9bdac9a1) in parent group-v264556. [ 875.415038] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Creating folder: Instances. Parent ref: group-v264661. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 875.415038] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 875.415038] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e1393f-eb49-3684-48f8-2093e52c3b1c" [ 875.415038] env[61947]: _type = "HttpNfcLease" [ 875.415038] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 875.415038] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f14467a-a5f1-49b0-86f7-3f0ded5ef020 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.419186] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5005e9e9-1c35-4bfd-ac6f-41288a9215c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.421683] env[61947]: DEBUG oslo_vmware.rw_handles [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520dd8da-4372-e989-4353-4482746d12cb/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 875.421947] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Uploaded image 13296250-6feb-41b0-b80f-26c6dd557dde to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 875.424550] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 875.430458] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8a45c9d1-abb7-4d47-8112-9efe823600ed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.432551] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 875.432551] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 875.433822] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Created folder: Instances in parent group-v264661. [ 875.435034] env[61947]: DEBUG oslo.service.loopingcall [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.435034] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.435034] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17a8ea6e-40f0-48e3-a051-ec2b7b29c106 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.456030] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 875.456030] env[61947]: value = "task-1224501" [ 875.456030] env[61947]: _type = "Task" [ 875.456030] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.516143] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.516143] env[61947]: value = "task-1224502" [ 875.516143] env[61947]: _type = "Task" [ 875.516143] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.523528] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224501, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.529195] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224502, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.545919] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ff54b61e-956e-4b06-a5e7-3b5bff11c482 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.647255] env[61947]: DEBUG nova.network.neutron [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.807578] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-716aa91e-e358-4402-aaf4-af4c1d6bc477 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.821946] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38073d7d-9a40-408e-a7e1-0731f6a6047a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.867406] env[61947]: DEBUG nova.compute.manager [req-394d113f-b355-4794-bbb6-30081405dfc5 req-87f0694e-594e-4778-8d28-9fafbc3bc878 service nova] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Detach interface failed, port_id=7d722a23-d40a-41cc-8e2a-c4456d5b456c, reason: Instance ee3b9509-5437-41b3-b612-91d148a338b6 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 876.033685] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224501, 'name': Destroy_Task, 'duration_secs': 0.416142} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.033685] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Destroyed the VM [ 876.033989] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 876.037665] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0929d221-4762-416f-a9ab-2000b87a2980 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.039906] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224502, 'name': CreateVM_Task, 'duration_secs': 0.380969} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.040155] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.041354] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.041714] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.042526] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.042917] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa1ea876-2647-4a6e-83b2-02cd6fcb7da1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.049111] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 876.049111] env[61947]: value = "task-1224503" [ 876.049111] env[61947]: _type = "Task" [ 876.049111] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.054508] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 876.054508] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f8c4d7-4f64-7ff8-8f70-c72ff08d81f7" [ 876.054508] env[61947]: _type = "Task" [ 876.054508] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.065113] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224503, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.071260] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f8c4d7-4f64-7ff8-8f70-c72ff08d81f7, 'name': SearchDatastore_Task, 'duration_secs': 0.012278} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.078569] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.080558] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.080961] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.081249] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.081462] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.082917] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8d9995e-c7fc-4b9c-9730-1a5ae75e2024 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.096022] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.096022] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.096022] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb462385-b3c2-429e-ac4b-b71d746347c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.101873] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 876.101873] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a604cb-4632-63f0-f0b6-2777f250d974" [ 876.101873] env[61947]: _type = "Task" [ 876.101873] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.118782] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a604cb-4632-63f0-f0b6-2777f250d974, 'name': SearchDatastore_Task, 'duration_secs': 0.010779} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.119679] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bacd41c8-b428-4620-95ce-93cc2d4fda1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.130553] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 876.130553] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525ce502-cd0c-5cbd-073f-27c0fd81e148" [ 876.130553] env[61947]: _type = "Task" [ 876.130553] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.139346] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525ce502-cd0c-5cbd-073f-27c0fd81e148, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.150015] env[61947]: INFO nova.compute.manager [-] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Took 1.30 seconds to deallocate network for instance. [ 876.178841] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384e0bcf-c190-4e28-bdde-b058869ed2b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.192706] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4956356f-3474-4504-917f-5c05236a46c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.237127] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727642f1-275e-4178-95db-3795cb3a73a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.247759] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bef74e3-bce7-4978-a133-697568cf7f51 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.264205] env[61947]: DEBUG nova.compute.provider_tree [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.325729] env[61947]: DEBUG nova.network.neutron [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.561038] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224503, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.621962] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.622651] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.647505] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525ce502-cd0c-5cbd-073f-27c0fd81e148, 'name': SearchDatastore_Task, 'duration_secs': 0.011134} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.650416] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.650416] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 876.650416] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cb2d7aa-b417-43e9-8d9a-d38b8a24d9c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.661115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.667384] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 876.667384] env[61947]: value = "task-1224504" [ 876.667384] env[61947]: _type = "Task" [ 876.667384] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.685726] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.767960] env[61947]: DEBUG nova.scheduler.client.report [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.831637] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.062256] env[61947]: DEBUG oslo_vmware.api [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224503, 'name': RemoveSnapshot_Task, 'duration_secs': 0.873919} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.062696] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 877.063102] env[61947]: INFO nova.compute.manager [None req-076885c5-c167-420a-bf1e-270ba74e363c tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Took 15.83 seconds to snapshot the instance on the hypervisor. [ 877.125493] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.180607] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505105} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.180869] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 877.181203] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.181529] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7603a8d4-07fd-4f80-8769-b4f66dcda6ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.188373] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 877.188373] env[61947]: value = "task-1224505" [ 877.188373] env[61947]: _type = "Task" [ 877.188373] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.196927] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.276544] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.277940] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.809s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.655920] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.699121] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069518} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.699524] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.702564] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a809a34-f892-44d1-8aa9-8e9c3871a8d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.737134] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.737824] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ef9384b-1916-4ebf-b090-ebb57c865eb5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.761408] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 877.761408] env[61947]: value = "task-1224506" [ 877.761408] env[61947]: _type = "Task" [ 877.761408] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.774014] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.781330] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "99e81c0b-b536-4933-9ade-c1e3b5c46cb4" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.781688] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "99e81c0b-b536-4933-9ade-c1e3b5c46cb4" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.786842] env[61947]: INFO nova.compute.claims [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.959699] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 877.960955] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce362706-209e-43bb-bdbd-50a191f7a0a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.968942] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 877.969781] env[61947]: ERROR oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk due to incomplete transfer. [ 877.969781] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-56072868-1c34-42bd-a88c-618f4d7ddafb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.977379] env[61947]: DEBUG oslo_vmware.rw_handles [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528995ab-95da-ad25-8550-c9bdf4048cf6/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 877.978113] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Uploaded image ef8d583a-38fd-4d74-9ed7-f1718e80cace to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 877.979974] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 877.980316] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-142e2586-4f92-4de6-aaea-326bca29ca48 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.987431] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 877.987431] env[61947]: value = "task-1224507" [ 877.987431] env[61947]: _type = "Task" [ 877.987431] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.000704] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224507, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.271427] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.294181] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "99e81c0b-b536-4933-9ade-c1e3b5c46cb4" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.512s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.294881] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.299525] env[61947]: INFO nova.compute.resource_tracker [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating resource usage from migration 89e41994-bff5-4166-a6a9-10258b48d1d3 [ 878.353767] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df2134a-677c-45cf-8be7-0cd0bad9b2de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.383493] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 878.506840] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224507, 'name': Destroy_Task, 'duration_secs': 0.349275} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.506840] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Destroyed the VM [ 878.506840] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.506840] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-444dfe70-1104-467a-9f81-11d118cbb108 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.515544] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 878.515544] env[61947]: value = "task-1224511" [ 878.515544] env[61947]: _type = "Task" [ 878.515544] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.529178] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224511, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.629143] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc504154-91c1-448d-a3fe-4ba03001d2dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.637292] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798b612b-6ccc-44e0-865e-720cf66fb835 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.675997] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58cd9e0-d0a9-4189-81cf-faa5da93e87d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.685017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac81eb3-74f8-476f-b97e-a3f88561bd00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.702242] env[61947]: DEBUG nova.compute.provider_tree [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.775687] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224506, 'name': ReconfigVM_Task, 'duration_secs': 0.820879} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.776216] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Reconfigured VM instance instance-00000045 to attach disk [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.777559] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a43b074-f0bf-44fe-995b-e4483d2e9b7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.785278] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 878.785278] env[61947]: value = "task-1224512" [ 878.785278] env[61947]: _type = "Task" [ 878.785278] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.798914] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224512, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.804568] env[61947]: DEBUG nova.compute.utils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.806324] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.806637] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.896888] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 878.896888] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-073b9825-fd8e-4ace-ae56-d237dc099690 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.905680] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 878.905680] env[61947]: value = "task-1224513" [ 878.905680] env[61947]: _type = "Task" [ 878.905680] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.912971] env[61947]: DEBUG nova.policy [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91dbf31be16c4c7f890536f744efdafc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea4fabf7e2ae4fe499160eb4e44f6677', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 878.920502] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224513, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.026891] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224511, 'name': RemoveSnapshot_Task, 'duration_secs': 0.465588} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.027279] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 879.027605] env[61947]: DEBUG nova.compute.manager [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.028500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1aa47d-c235-4ccc-a939-7e5d45d828cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.209018] env[61947]: DEBUG nova.scheduler.client.report [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.301020] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224512, 'name': Rename_Task, 'duration_secs': 0.150031} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.301326] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 879.301619] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1587a8e4-14a8-49b0-909f-58ed081b7c8d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.309467] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 879.309467] env[61947]: value = "task-1224514" [ 879.309467] env[61947]: _type = "Task" [ 879.309467] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.309735] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.326974] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224514, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.416873] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224513, 'name': PowerOffVM_Task, 'duration_secs': 0.318158} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.417745] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 879.417745] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 879.535946] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Successfully created port: 87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.542628] env[61947]: INFO nova.compute.manager [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Shelve offloading [ 879.544540] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.544786] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-408ec4a8-ee3b-4db1-bd23-e93b0cda64eb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.554144] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 879.554144] env[61947]: value = "task-1224515" [ 879.554144] env[61947]: _type = "Task" [ 879.554144] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.562812] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 879.563052] env[61947]: DEBUG nova.compute.manager [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.563847] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcc1235-537f-4df7-90f4-46b134e621df {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.569945] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.570099] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.570284] env[61947]: DEBUG nova.network.neutron [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.718681] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.441s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.718982] env[61947]: INFO nova.compute.manager [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Migrating [ 879.726927] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.831s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.728510] env[61947]: INFO nova.compute.claims [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.824399] env[61947]: DEBUG oslo_vmware.api [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224514, 'name': PowerOnVM_Task, 'duration_secs': 0.455371} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.824722] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 879.825591] env[61947]: INFO nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Took 9.55 seconds to spawn the instance on the hypervisor. [ 879.825933] env[61947]: DEBUG nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.827269] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32499147-d839-475f-adff-54e46075966b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.928178] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.935593] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-461ffd43-b2fd-4b08-b45a-9d4d8c84599a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.950422] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.950763] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.958137] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 879.958137] env[61947]: value = "task-1224516" [ 879.958137] env[61947]: _type = "Task" [ 879.958137] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.972963] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224516, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.249734] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.250915] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.250915] env[61947]: DEBUG nova.network.neutron [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.332897] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.356576] env[61947]: INFO nova.compute.manager [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Took 37.11 seconds to build instance. [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.380099] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.380533] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.380533] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.380760] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.380891] env[61947]: DEBUG nova.virt.hardware [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.381847] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51a3703-0943-40a3-85e1-cd5536328f88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.398799] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08e777a-46ef-4499-937a-b8af99dac9b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.454340] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.469633] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224516, 'name': ReconfigVM_Task, 'duration_secs': 0.246462} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.469977] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 880.571023] env[61947]: DEBUG nova.network.neutron [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.867554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-baf13f57-a684-4330-b4c0-c7a405ca6913 tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.636s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.976462] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.976804] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.977023] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.977266] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.977449] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.977652] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.978063] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.980122] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.980122] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.980122] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.980122] env[61947]: DEBUG nova.virt.hardware [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.987706] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 880.992503] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c1814c4-d862-45ba-aded-9e3c5942b054 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.013611] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.020500] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 881.020500] env[61947]: value = "task-1224518" [ 881.020500] env[61947]: _type = "Task" [ 881.020500] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.035867] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224518, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.079081] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.142410] env[61947]: DEBUG nova.network.neutron [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.157028] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdf4837-5656-4f6c-b20a-f78a5c46c50b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.169930] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ce4cfb-7188-446a-8922-dcf672d5a59d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.214369] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df60cbc-448d-4c33-9261-51ab18dc4654 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.222534] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d352100-8153-4d13-9f9b-cf0d0ff2e541 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.238753] env[61947]: DEBUG nova.compute.provider_tree [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.366061] env[61947]: DEBUG nova.compute.manager [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Received event network-vif-plugged-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.366449] env[61947]: DEBUG oslo_concurrency.lockutils [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] Acquiring lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.366534] env[61947]: DEBUG oslo_concurrency.lockutils [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.366786] env[61947]: DEBUG oslo_concurrency.lockutils [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.366856] env[61947]: DEBUG nova.compute.manager [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] No waiting events found dispatching network-vif-plugged-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.367080] env[61947]: WARNING nova.compute.manager [req-5d9e778d-8b22-44dc-af17-91f200cdfb1d req-af129cb3-ad55-471e-83f3-e34346e6ee18 service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Received unexpected event network-vif-plugged-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b for instance with vm_state building and task_state spawning. [ 881.432151] env[61947]: DEBUG nova.compute.manager [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-vif-unplugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.432616] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.432891] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.433064] env[61947]: DEBUG oslo_concurrency.lockutils [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.433318] env[61947]: DEBUG nova.compute.manager [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] No waiting events found dispatching network-vif-unplugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.433430] env[61947]: WARNING nova.compute.manager [req-c3f9b58e-bf1a-4c5f-b116-7799034d18a8 req-a5a3dcb0-1082-41f0-955c-8f1fa71c1e2b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received unexpected event network-vif-unplugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 for instance with vm_state shelved and task_state shelving_offloading. [ 881.509406] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.511639] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e92d6b-d11f-4f35-9ec7-c23593afa44d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.522893] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 881.526810] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2057f53-a927-4cc1-b904-e2488f5e3277 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.535784] env[61947]: INFO nova.compute.manager [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Rebuilding instance [ 881.539028] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224518, 'name': ReconfigVM_Task, 'duration_secs': 0.217064} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.539699] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 881.544208] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a4e998-969a-47ec-bb62-caa7f7b35661 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.569145] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.569936] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Successfully updated port: 87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.576315] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4e80abc-adca-4907-a89b-5815836e69f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.607221] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 881.607221] env[61947]: value = "task-1224520" [ 881.607221] env[61947]: _type = "Task" [ 881.607221] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.612529] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 881.612856] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 881.612977] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleting the datastore file [datastore2] fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.613698] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11aff7b5-d526-4711-a76b-2c4ed170f8c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.618934] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224520, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.626402] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 881.626402] env[61947]: value = "task-1224521" [ 881.626402] env[61947]: _type = "Task" [ 881.626402] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.632286] env[61947]: DEBUG nova.compute.manager [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.633068] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8781a40-6e2f-44b7-980a-ed1275fafe3f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.641211] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224521, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.645500] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.744262] env[61947]: DEBUG nova.scheduler.client.report [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.095497] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.095866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquired lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.095866] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.117197] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.138320] env[61947]: DEBUG oslo_vmware.api [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224521, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262057} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.138664] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.138887] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 882.139101] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 882.148624] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 882.152672] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-996e5aee-67a8-4853-b0d0-28e3a2e8239d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.159343] env[61947]: INFO nova.scheduler.client.report [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted allocations for instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 [ 882.162062] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 882.162062] env[61947]: value = "task-1224522" [ 882.162062] env[61947]: _type = "Task" [ 882.162062] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.174183] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.249079] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.249835] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.254110] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.125s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.255676] env[61947]: INFO nova.compute.claims [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.553982] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.553982] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.618323] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224520, 'name': ReconfigVM_Task, 'duration_secs': 0.95152} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.618676] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Reconfigured VM instance instance-0000003d to attach disk [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd/eede4f34-5ac4-475c-a74b-a98327f648fd.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.618949] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 882.632487] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.663841] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.674919] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224522, 'name': PowerOffVM_Task, 'duration_secs': 0.130992} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.677371] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.677371] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 882.677371] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7a0d10-f495-4e94-809d-2289d6493b95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.685773] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.686700] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ee68fb8-86fe-421a-8429-1964fc55637b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.718567] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 882.718853] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 882.719136] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Deleting the datastore file [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.719439] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2a4593b-d50b-491f-bf29-ef55477236a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.726233] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 882.726233] env[61947]: value = "task-1224524" [ 882.726233] env[61947]: _type = "Task" [ 882.726233] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.734974] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.760336] env[61947]: DEBUG nova.compute.utils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.763562] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.763694] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.809640] env[61947]: DEBUG nova.network.neutron [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Updating instance_info_cache with network_info: [{"id": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "address": "fa:16:3e:fc:51:ca", "network": {"id": "16ddccb3-c3d1-4b8b-8af0-eefb5ff6cfa9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977624959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea4fabf7e2ae4fe499160eb4e44f6677", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e10fb3-d4", "ovs_interfaceid": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.009128] env[61947]: DEBUG nova.policy [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd1ffc62b391469da9bbdfa2763e07e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '025e8ea7833e47e49562ef9ea3bb50a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 883.058656] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.129112] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0dc5f0-c72e-437f-bdd0-e48aa1bb8eaa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.150313] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fdf1d4-f8be-4466-bedd-4d70812c32b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.170906] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 883.178250] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb92d52-e1e4-456b-887e-a57e0f743ce1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.194325] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 883.236381] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10298} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.236666] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.236864] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 883.237060] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 883.269379] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.311654] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Releasing lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.312069] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Instance network_info: |[{"id": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "address": "fa:16:3e:fc:51:ca", "network": {"id": "16ddccb3-c3d1-4b8b-8af0-eefb5ff6cfa9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977624959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea4fabf7e2ae4fe499160eb4e44f6677", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e10fb3-d4", "ovs_interfaceid": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.312971] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:51:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.321981] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Creating folder: Project (ea4fabf7e2ae4fe499160eb4e44f6677). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.322755] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78a72306-74fe-41ac-a4fa-e2bd8504e7a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.338903] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Created folder: Project (ea4fabf7e2ae4fe499160eb4e44f6677) in parent group-v264556. [ 883.339253] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Creating folder: Instances. Parent ref: group-v264667. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.342758] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Successfully created port: 73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.344280] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65af6612-adca-4df0-8ddf-be8708bd0c8d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.355056] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Created folder: Instances in parent group-v264667. [ 883.355056] env[61947]: DEBUG oslo.service.loopingcall [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.355056] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.355531] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f943a9d-467b-4ade-87e9-8d8b209b2237 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.380290] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.380290] env[61947]: value = "task-1224528" [ 883.380290] env[61947]: _type = "Task" [ 883.380290] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.390202] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224528, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.580497] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.604603] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f31952-9367-437b-aa86-c9bdbbfc65bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.614240] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394f1ff6-3350-4a8e-9bf6-842c137b100a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.646652] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Successfully created port: 6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.649357] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7734f81-2238-401b-9384-5905ec1725a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.658152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe2a5fa-c07c-4f11-8bee-63e6bd7ca251 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.673746] env[61947]: DEBUG nova.compute.provider_tree [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.691128] env[61947]: DEBUG nova.compute.manager [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Received event network-changed-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.691128] env[61947]: DEBUG nova.compute.manager [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Refreshing instance network info cache due to event network-changed-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.691128] env[61947]: DEBUG oslo_concurrency.lockutils [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] Acquiring lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.691128] env[61947]: DEBUG oslo_concurrency.lockutils [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] Acquired lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.691448] env[61947]: DEBUG nova.network.neutron [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Refreshing network info cache for port 87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.700326] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.700456] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-309e61a2-851e-4de7-9c77-fda7e0cef95e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.708032] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 883.708032] env[61947]: value = "task-1224529" [ 883.708032] env[61947]: _type = "Task" [ 883.708032] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.713968] env[61947]: DEBUG nova.compute.manager [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.713968] env[61947]: DEBUG nova.compute.manager [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing instance network info cache due to event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.714225] env[61947]: DEBUG oslo_concurrency.lockutils [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.714396] env[61947]: DEBUG oslo_concurrency.lockutils [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.714577] env[61947]: DEBUG nova.network.neutron [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.720451] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.724822] env[61947]: DEBUG nova.network.neutron [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Port 5e37697a-486f-4961-aed8-8ccdbf59992b binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 883.816123] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 883.817213] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d084147a-f9d2-4893-bc26-ea6db2024d75 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.825983] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 883.826212] env[61947]: ERROR oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk due to incomplete transfer. [ 883.826453] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-81e0f178-2b7b-4a45-b8c7-ba36488bd197 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.833489] env[61947]: DEBUG oslo_vmware.rw_handles [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dfa77d-7b9d-9bf3-70fb-6618d99b48bb/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 883.833741] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploaded image 2cd5eb90-64f8-4008-92f1-e73346f0afac to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 883.836506] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 883.836884] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-67366f68-330a-4a15-8ef2-1d3fd2d8d1af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.843300] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 883.843300] env[61947]: value = "task-1224530" [ 883.843300] env[61947]: _type = "Task" [ 883.843300] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.853200] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224530, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.891189] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224528, 'name': CreateVM_Task, 'duration_secs': 0.342016} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.891394] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 883.892725] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.892725] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.893031] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.893103] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fadafd37-1861-4396-b825-84714682c917 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.898575] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 883.898575] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5275752b-b563-0f92-3e64-bfc02d4a0054" [ 883.898575] env[61947]: _type = "Task" [ 883.898575] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.907487] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5275752b-b563-0f92-3e64-bfc02d4a0054, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.177409] env[61947]: DEBUG nova.scheduler.client.report [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.221311] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224529, 'name': PowerOffVM_Task, 'duration_secs': 0.194423} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.221633] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.221858] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 884.276733] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.277274] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.277474] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.277770] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.277951] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.278188] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.278486] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.278673] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.278951] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.279114] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.279597] env[61947]: DEBUG nova.virt.hardware [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.281221] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.284011] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86802963-7b2c-422b-bdb5-7ff6df7bd9e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.296891] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697da3fc-4e6b-4dc2-9b9e-21b5dc08fb7e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.315149] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance VIF info [] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.321032] env[61947]: DEBUG oslo.service.loopingcall [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.325578] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 884.326024] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e52db06-2f1c-4e91-9569-8e635f72442b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.343048] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.343343] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.343510] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.343700] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.344139] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.344334] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.344555] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.344724] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.345095] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.345311] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.345566] env[61947]: DEBUG nova.virt.hardware [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.346765] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4e8514-f46f-4548-a151-db69d03dbcf2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.358754] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.358754] env[61947]: value = "task-1224531" [ 884.358754] env[61947]: _type = "Task" [ 884.358754] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.372794] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc519f1-a813-4981-92b4-8364d778f7f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.377017] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224530, 'name': Destroy_Task, 'duration_secs': 0.32184} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.377204] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224531, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.379380] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroyed the VM [ 884.379669] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 884.380508] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5eb6fa1e-ce52-4df3-a280-3a64b71a3de3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.393694] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 884.393694] env[61947]: value = "task-1224532" [ 884.393694] env[61947]: _type = "Task" [ 884.393694] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.405267] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224532, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.414413] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5275752b-b563-0f92-3e64-bfc02d4a0054, 'name': SearchDatastore_Task, 'duration_secs': 0.00943} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.414660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.414961] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.415176] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.415336] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.415635] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.415921] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f75e7dae-d589-432e-a9ca-7541e71d546c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.423589] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.423768] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 884.424531] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f06e1fa-a69a-45f0-bfe1-b3413b3efc1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.430019] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 884.430019] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524220a4-fb7c-911e-bddf-fa0a7d508d4d" [ 884.430019] env[61947]: _type = "Task" [ 884.430019] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.439928] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524220a4-fb7c-911e-bddf-fa0a7d508d4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.561201] env[61947]: DEBUG nova.network.neutron [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updated VIF entry in instance network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.561924] env[61947]: DEBUG nova.network.neutron [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.584183] env[61947]: DEBUG nova.network.neutron [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Updated VIF entry in instance network info cache for port 87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.584543] env[61947]: DEBUG nova.network.neutron [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Updating instance_info_cache with network_info: [{"id": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "address": "fa:16:3e:fc:51:ca", "network": {"id": "16ddccb3-c3d1-4b8b-8af0-eefb5ff6cfa9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977624959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea4fabf7e2ae4fe499160eb4e44f6677", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e10fb3-d4", "ovs_interfaceid": "87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.685050] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.685412] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.688018] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.678s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.688257] env[61947]: DEBUG nova.objects.instance [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lazy-loading 'resources' on Instance uuid 4f0b9636-4793-4b1c-8b2b-7f31a15669bb {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.734286] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.734562] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.734730] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.734915] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.735075] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.735229] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.735431] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.735596] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.735763] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.736085] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.736148] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.751783] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-216a41f9-1324-4d2e-b780-800367e09e5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.766799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.767070] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.767260] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.775090] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 884.775090] env[61947]: value = "task-1224533" [ 884.775090] env[61947]: _type = "Task" [ 884.775090] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.785473] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.868314] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224531, 'name': CreateVM_Task, 'duration_secs': 0.297528} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.868684] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.868977] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.869167] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.871020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.871020] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90efb99c-d9ba-4e84-804b-3d2274ee470a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.874252] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 884.874252] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fe6453-39ca-3c2e-7462-f5178137edfd" [ 884.874252] env[61947]: _type = "Task" [ 884.874252] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.884275] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fe6453-39ca-3c2e-7462-f5178137edfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.902720] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224532, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.940600] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524220a4-fb7c-911e-bddf-fa0a7d508d4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008661} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.941499] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9b30fd4-661c-4c83-b4a4-863308dce4b9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.947065] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 884.947065] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c143b9-792e-0bcf-38fe-ca4dd8bef938" [ 884.947065] env[61947]: _type = "Task" [ 884.947065] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.955432] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c143b9-792e-0bcf-38fe-ca4dd8bef938, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.998250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.065102] env[61947]: DEBUG oslo_concurrency.lockutils [req-3af75560-be4d-4ad9-8e36-c8cbcb051ec4 req-990d3a66-61ec-4912-a1ad-cc752014f18b service nova] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.088321] env[61947]: DEBUG oslo_concurrency.lockutils [req-e555940e-bca4-4bc0-b11c-1570b7feec6c req-e355c2d9-2059-4d70-a116-f1b78c3c3c8f service nova] Releasing lock "refresh_cache-7f308223-6c3c-4c62-8401-d7a44eeb69d2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.192284] env[61947]: DEBUG nova.compute.utils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.197908] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Successfully updated port: 73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.199859] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.200087] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.269913] env[61947]: DEBUG nova.policy [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 885.286869] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224533, 'name': ReconfigVM_Task, 'duration_secs': 0.309757} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.290138] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 885.385830] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fe6453-39ca-3c2e-7462-f5178137edfd, 'name': SearchDatastore_Task, 'duration_secs': 0.009278} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.388116] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.388477] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.388752] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.404474] env[61947]: DEBUG oslo_vmware.api [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224532, 'name': RemoveSnapshot_Task, 'duration_secs': 0.596826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.407329] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 885.407626] env[61947]: INFO nova.compute.manager [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 15.37 seconds to snapshot the instance on the hypervisor. [ 885.459776] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c143b9-792e-0bcf-38fe-ca4dd8bef938, 'name': SearchDatastore_Task, 'duration_secs': 0.009677} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.460064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.460342] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 7f308223-6c3c-4c62-8401-d7a44eeb69d2/7f308223-6c3c-4c62-8401-d7a44eeb69d2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 885.460628] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.460836] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.461129] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-006efc97-f885-4b38-b63f-edb87364b4d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.463228] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d37aa095-e2e5-4d21-bdf1-e5d0253ec0d4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.469882] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 885.469882] env[61947]: value = "task-1224535" [ 885.469882] env[61947]: _type = "Task" [ 885.469882] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.474010] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.474242] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.475674] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-991bb936-9136-4f1e-b349-faa7f5275356 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.481721] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd17e899-9bb1-4528-9d12-241b52e71d4d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.486624] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.493069] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 885.493069] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528f00bd-c8ce-2904-920c-d8d31feca1b1" [ 885.493069] env[61947]: _type = "Task" [ 885.493069] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.494274] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5267a8-ae85-4df6-842c-004a9094e918 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.510125] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528f00bd-c8ce-2904-920c-d8d31feca1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.007813} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.544102] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56598e8c-9f30-4741-8c4d-89a1822ad30a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.547628] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b55837-e6cf-4a8e-ac6d-185cdd580e36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.554022] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 885.554022] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528d05f0-1d25-75bd-ec83-bb489399d59a" [ 885.554022] env[61947]: _type = "Task" [ 885.554022] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.560734] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b252ac9b-0d6b-4a6d-ab17-cb201d286a9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.571353] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528d05f0-1d25-75bd-ec83-bb489399d59a, 'name': SearchDatastore_Task, 'duration_secs': 0.008424} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.578178] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.578455] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 885.578968] env[61947]: DEBUG nova.compute.provider_tree [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.580727] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb08092e-3c9a-469c-a67e-d429cd981c8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.591227] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 885.591227] env[61947]: value = "task-1224536" [ 885.591227] env[61947]: _type = "Task" [ 885.591227] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.600483] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.632403] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Successfully created port: dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.702581] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.796587] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:48:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='9a82a007-fbf5-4f42-9bd3-1a8156f71854',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-569079550',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.796892] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.797170] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.797432] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.797625] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.797825] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.798178] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.798449] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.798712] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.798989] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.799289] env[61947]: DEBUG nova.virt.hardware [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.805231] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 885.806027] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-185e6385-e83d-42b7-b1a4-3d7542b42831 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.831638] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 885.831638] env[61947]: value = "task-1224537" [ 885.831638] env[61947]: _type = "Task" [ 885.831638] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.845424] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224537, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.846920] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.847843] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.848278] env[61947]: DEBUG nova.network.neutron [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.984854] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224535, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470687} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.987419] env[61947]: DEBUG nova.compute.manager [None req-c7686d97-67fd-483a-9d0a-f838caae5278 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Found 2 images (rotation: 2) {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 885.988704] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 7f308223-6c3c-4c62-8401-d7a44eeb69d2/7f308223-6c3c-4c62-8401-d7a44eeb69d2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.989099] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.989693] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da4fecee-9f0c-4b37-ae2d-c271119a84ce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.005028] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 886.005028] env[61947]: value = "task-1224538" [ 886.005028] env[61947]: _type = "Task" [ 886.005028] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.019314] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.037453] env[61947]: DEBUG nova.compute.manager [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-vif-plugged-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.037453] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Acquiring lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.037453] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.037453] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.037453] env[61947]: DEBUG nova.compute.manager [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] No waiting events found dispatching network-vif-plugged-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.037453] env[61947]: WARNING nova.compute.manager [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received unexpected event network-vif-plugged-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 for instance with vm_state building and task_state spawning. [ 886.037453] env[61947]: DEBUG nova.compute.manager [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-changed-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.038087] env[61947]: DEBUG nova.compute.manager [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Refreshing instance network info cache due to event network-changed-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.038423] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Acquiring lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.038714] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Acquired lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.038992] env[61947]: DEBUG nova.network.neutron [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Refreshing network info cache for port 73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.087044] env[61947]: DEBUG nova.scheduler.client.report [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.104795] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224536, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.342322] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224537, 'name': ReconfigVM_Task, 'duration_secs': 0.175694} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.342322] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 886.343412] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9180727c-b4c8-422e-8c41-6ebf180d374f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.371187] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.371625] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c2cc3c4-7087-4f92-9117-529f94b95494 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.395012] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 886.395012] env[61947]: value = "task-1224539" [ 886.395012] env[61947]: _type = "Task" [ 886.395012] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.404076] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224539, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.513720] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087204} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.514081] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.514979] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bea974-e5ff-447b-ba19-8942d280a45f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.539395] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 7f308223-6c3c-4c62-8401-d7a44eeb69d2/7f308223-6c3c-4c62-8401-d7a44eeb69d2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.541924] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1495a9a-6cae-4424-ab52-62ce03c3a64b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.564789] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 886.564789] env[61947]: value = "task-1224540" [ 886.564789] env[61947]: _type = "Task" [ 886.564789] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.574019] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224540, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.588483] env[61947]: DEBUG nova.network.neutron [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.591156] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.593968] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.933s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.593968] env[61947]: DEBUG nova.objects.instance [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lazy-loading 'resources' on Instance uuid ee3b9509-5437-41b3-b612-91d148a338b6 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.605292] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692841} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.605292] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 886.605292] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.605292] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e8863ea-0c86-4047-8170-059a3908ce56 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.614446] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 886.614446] env[61947]: value = "task-1224541" [ 886.614446] env[61947]: _type = "Task" [ 886.614446] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.626696] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.636291] env[61947]: INFO nova.scheduler.client.report [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Deleted allocations for instance 4f0b9636-4793-4b1c-8b2b-7f31a15669bb [ 886.692022] env[61947]: DEBUG nova.network.neutron [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.718203] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.721091] env[61947]: DEBUG nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.722364] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc92c09-4f2f-410c-aa37-29ef51e65dbd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.752050] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.752381] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.752550] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.752959] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.753207] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.753313] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.753530] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.753698] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.753905] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.754045] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.754232] env[61947]: DEBUG nova.virt.hardware [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.755148] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03067035-c000-43fe-9100-918bead52dc2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.763878] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04217d11-9b19-4eda-8607-2c243e50e2be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.790632] env[61947]: DEBUG nova.network.neutron [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.904953] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224539, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.075702] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.124863] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075942} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.125395] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.126181] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdab9c0-6930-41dc-8686-9a9b3bdd675f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.166683] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.173301] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7be900a4-ae05-4883-8de9-a744b06de14a tempest-ServerPasswordTestJSON-1269963253 tempest-ServerPasswordTestJSON-1269963253-project-member] Lock "4f0b9636-4793-4b1c-8b2b-7f31a15669bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.087s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.173301] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49f0a95d-6df4-4c26-bd11-87b9334d2b25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.201302] env[61947]: DEBUG oslo_concurrency.lockutils [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.213859] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 887.213859] env[61947]: value = "task-1224543" [ 887.213859] env[61947]: _type = "Task" [ 887.213859] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.226953] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224543, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.236583] env[61947]: INFO nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] instance snapshotting [ 887.237212] env[61947]: DEBUG nova.objects.instance [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.294943] env[61947]: DEBUG oslo_concurrency.lockutils [req-399ab0a4-50bb-4d61-8aa2-62a0fc7ec024 req-7c79e86f-dca8-4a34-b7c6-2db3182e7f69 service nova] Releasing lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.377227] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Successfully updated port: dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.405992] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224539, 'name': ReconfigVM_Task, 'duration_secs': 0.800416} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.408925] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.409255] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 887.491435] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d58046-7404-43be-bc56-2657552d5551 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.504379] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c9255a-67d3-4783-934b-605d6d6607a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.542560] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392b8045-4286-4516-a2c4-ac4dcd8d7fef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.555556] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f86e2b-9547-4827-8cdf-ba5a4a45f5ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.574524] env[61947]: DEBUG nova.compute.provider_tree [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.589250] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224540, 'name': ReconfigVM_Task, 'duration_secs': 0.626398} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.589589] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 7f308223-6c3c-4c62-8401-d7a44eeb69d2/7f308223-6c3c-4c62-8401-d7a44eeb69d2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.590491] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce4e3e94-f0c2-4554-b8ea-b794973a5196 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.602627] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 887.602627] env[61947]: value = "task-1224544" [ 887.602627] env[61947]: _type = "Task" [ 887.602627] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.619821] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224544, 'name': Rename_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.720288] env[61947]: DEBUG nova.compute.manager [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-vif-plugged-6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.720373] env[61947]: DEBUG oslo_concurrency.lockutils [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] Acquiring lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.720708] env[61947]: DEBUG oslo_concurrency.lockutils [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.720914] env[61947]: DEBUG oslo_concurrency.lockutils [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.720914] env[61947]: DEBUG nova.compute.manager [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] No waiting events found dispatching network-vif-plugged-6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.721064] env[61947]: WARNING nova.compute.manager [req-3db69257-884a-49c8-b7ee-5ca03cd8142f req-b2b6859c-25d6-4976-ba1c-9344c149552d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received unexpected event network-vif-plugged-6832c702-f2d4-446f-a6e3-9225cf51f491 for instance with vm_state building and task_state spawning. [ 887.733755] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.743656] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7396ffc5-b952-4ca4-a837-e8c32de70e2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.749784] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee07d11-1309-43ac-8da9-5b1ef6efd601 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.767875] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dccf7f-5be5-423f-9724-2318912415e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.785528] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63e9aca-9535-4b7a-8097-39c07f6497bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.799506] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 887.826549] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Successfully updated port: 6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.881839] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.881936] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.882066] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.919042] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9164165-7538-404e-adaf-e5ce7179191d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.939211] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222b171a-a04f-46d3-b39a-2bcb332b9bc9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.962694] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 888.064318] env[61947]: DEBUG nova.compute.manager [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received event network-vif-plugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.064551] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.065150] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.065338] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.065513] env[61947]: DEBUG nova.compute.manager [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] No waiting events found dispatching network-vif-plugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.066181] env[61947]: WARNING nova.compute.manager [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received unexpected event network-vif-plugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef for instance with vm_state building and task_state spawning. [ 888.066181] env[61947]: DEBUG nova.compute.manager [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received event network-changed-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.066181] env[61947]: DEBUG nova.compute.manager [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Refreshing instance network info cache due to event network-changed-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 888.066378] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Acquiring lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.082036] env[61947]: DEBUG nova.scheduler.client.report [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.114175] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224544, 'name': Rename_Task, 'duration_secs': 0.216679} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.114520] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.114799] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bee204f8-2dd0-4207-9b70-caa2d1a85fba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.124402] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 888.124402] env[61947]: value = "task-1224545" [ 888.124402] env[61947]: _type = "Task" [ 888.124402] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.138620] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224545, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.228482] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224543, 'name': ReconfigVM_Task, 'duration_secs': 0.705473} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.228854] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Reconfigured VM instance instance-00000045 to attach disk [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2/e0eebf59-c84e-4462-b280-d783a04525e2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.229433] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de3ce280-d903-41c7-bbbf-a69e9b9cbe36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.237278] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 888.237278] env[61947]: value = "task-1224546" [ 888.237278] env[61947]: _type = "Task" [ 888.237278] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.245675] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224546, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.295289] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 888.295776] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3f648284-d62b-4ed0-86a3-54cd4f4fd0fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.303211] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.305034] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a460fb7-c2b1-4764-9cd3-16d80575d973 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.306905] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 888.306905] env[61947]: value = "task-1224547" [ 888.306905] env[61947]: _type = "Task" [ 888.306905] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.312557] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 888.312557] env[61947]: value = "task-1224548" [ 888.312557] env[61947]: _type = "Task" [ 888.312557] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.315700] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224547, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.325276] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.332230] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.332404] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.333020] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.477308] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.521273] env[61947]: DEBUG nova.network.neutron [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Port c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 888.586641] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.993s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.589236] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.933s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.591428] env[61947]: INFO nova.compute.claims [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.624225] env[61947]: INFO nova.scheduler.client.report [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Deleted allocations for instance ee3b9509-5437-41b3-b612-91d148a338b6 [ 888.646269] env[61947]: DEBUG oslo_vmware.api [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224545, 'name': PowerOnVM_Task, 'duration_secs': 0.510012} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.646406] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 888.646689] env[61947]: INFO nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Took 8.31 seconds to spawn the instance on the hypervisor. [ 888.646902] env[61947]: DEBUG nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.647752] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2462fe6d-482c-4194-8c51-81243c57ace3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.752028] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224546, 'name': Rename_Task, 'duration_secs': 0.293905} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.752792] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 888.752925] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19789192-7bab-450d-bb87-af57175c6fc2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.761204] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 888.761204] env[61947]: value = "task-1224549" [ 888.761204] env[61947]: _type = "Task" [ 888.761204] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.776080] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224549, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.821249] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224547, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.834415] env[61947]: DEBUG oslo_vmware.api [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224548, 'name': PowerOnVM_Task, 'duration_secs': 0.471873} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.834415] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 888.834415] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-58620884-0ca8-4c73-8878-50d21d553c80 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance 'eede4f34-5ac4-475c-a74b-a98327f648fd' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 888.847197] env[61947]: DEBUG nova.network.neutron [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updating instance_info_cache with network_info: [{"id": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "address": "fa:16:3e:18:eb:2c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcfe6b8e-1a", "ovs_interfaceid": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.885316] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.143472] env[61947]: DEBUG oslo_concurrency.lockutils [None req-912c05c3-c41b-4fa7-b301-20b925bcf341 tempest-AttachInterfacesV270Test-1748578177 tempest-AttachInterfacesV270Test-1748578177-project-member] Lock "ee3b9509-5437-41b3-b612-91d148a338b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.492s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.181675] env[61947]: INFO nova.compute.manager [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Took 35.33 seconds to build instance. [ 889.273094] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224549, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.320254] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224547, 'name': CreateSnapshot_Task, 'duration_secs': 0.968977} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.320254] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 889.320430] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fa6a25-3bb6-4e1a-97e7-99284d37d0e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.349047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.350627] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance network_info: |[{"id": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "address": "fa:16:3e:18:eb:2c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcfe6b8e-1a", "ovs_interfaceid": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.350627] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Acquired lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.351117] env[61947]: DEBUG nova.network.neutron [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Refreshing network info cache for port dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.352197] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:eb:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dcfe6b8e-1a14-4344-a03c-1f62c9b7baef', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.362385] env[61947]: DEBUG oslo.service.loopingcall [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.363642] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.363642] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bec9f473-323e-4b19-aa22-8cdc22ea06cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.384350] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.384350] env[61947]: value = "task-1224551" [ 889.384350] env[61947]: _type = "Task" [ 889.384350] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.392570] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224551, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.448588] env[61947]: DEBUG nova.network.neutron [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updating instance_info_cache with network_info: [{"id": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "address": "fa:16:3e:6e:4d:bc", "network": {"id": "b3f69232-c6a8-46da-a6c6-d9d3ec2ae81a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-522940052", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.179", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b59bc8-bc", "ovs_interfaceid": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6832c702-f2d4-446f-a6e3-9225cf51f491", "address": "fa:16:3e:0f:df:17", "network": {"id": "afbf9dc1-7cc0-4ce3-82f4-63d7f45a7950", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-514383215", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832c702-f2", "ovs_interfaceid": "6832c702-f2d4-446f-a6e3-9225cf51f491", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.553712] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.554094] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.554156] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.685778] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.686198] env[61947]: DEBUG oslo_concurrency.lockutils [None req-366fa52a-73ec-4205-a243-01bcd5a3d3e2 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.849s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.686678] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.686902] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.687195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.687376] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.690168] env[61947]: INFO nova.compute.manager [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Terminating instance [ 889.693890] env[61947]: DEBUG nova.compute.manager [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.694803] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.695575] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52931c34-c5cc-4211-b590-c2e0cae629db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.704981] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.707997] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26d49930-996b-4fa6-8f52-58319174b90c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.718285] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 889.718285] env[61947]: value = "task-1224552" [ 889.718285] env[61947]: _type = "Task" [ 889.718285] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.727964] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224552, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.773714] env[61947]: DEBUG oslo_vmware.api [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224549, 'name': PowerOnVM_Task, 'duration_secs': 0.81605} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.777491] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.777748] env[61947]: DEBUG nova.compute.manager [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.778883] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd710475-9921-47e6-9768-4225306067a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.851805] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 889.861630] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-91175a98-5b11-4d60-8012-434474f73bd2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.875052] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 889.875052] env[61947]: value = "task-1224553" [ 889.875052] env[61947]: _type = "Task" [ 889.875052] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.894832] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224553, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.904349] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224551, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.942296] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0639281-2cea-4fc9-a324-8ef373b46a35 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.951525] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.951525] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance network_info: |[{"id": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "address": "fa:16:3e:6e:4d:bc", "network": {"id": "b3f69232-c6a8-46da-a6c6-d9d3ec2ae81a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-522940052", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.179", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b59bc8-bc", "ovs_interfaceid": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6832c702-f2d4-446f-a6e3-9225cf51f491", "address": "fa:16:3e:0f:df:17", "network": {"id": "afbf9dc1-7cc0-4ce3-82f4-63d7f45a7950", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-514383215", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832c702-f2", "ovs_interfaceid": "6832c702-f2d4-446f-a6e3-9225cf51f491", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.952256] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:4d:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea45c024-d603-4bac-9c1b-f302437ea4fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:df:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97113f46-d648-4613-b233-069acba18198', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6832c702-f2d4-446f-a6e3-9225cf51f491', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.962381] env[61947]: DEBUG oslo.service.loopingcall [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.963398] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede74b5c-1b4d-4fd7-9179-4ba1211bec96 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.967178] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.967269] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fee36682-5ec9-4bc0-b747-c88a89aba15a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.023694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fc1f5a-d60d-4489-a258-e9941d105300 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.026656] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.026656] env[61947]: value = "task-1224554" [ 890.026656] env[61947]: _type = "Task" [ 890.026656] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.040024] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd229842-1ab0-4aff-a561-a102532f36ce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.041737] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224554, 'name': CreateVM_Task} progress is 15%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.054163] env[61947]: DEBUG nova.compute.provider_tree [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.193758] env[61947]: DEBUG nova.compute.manager [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-changed-6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.194107] env[61947]: DEBUG nova.compute.manager [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Refreshing instance network info cache due to event network-changed-6832c702-f2d4-446f-a6e3-9225cf51f491. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 890.194511] env[61947]: DEBUG oslo_concurrency.lockutils [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] Acquiring lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.194828] env[61947]: DEBUG oslo_concurrency.lockutils [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] Acquired lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.195243] env[61947]: DEBUG nova.network.neutron [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Refreshing network info cache for port 6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.231846] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224552, 'name': PowerOffVM_Task, 'duration_secs': 0.189262} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.231846] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.231846] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.232035] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69df52db-0fa1-425a-98ee-402ce77817f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.246613] env[61947]: DEBUG nova.network.neutron [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updated VIF entry in instance network info cache for port dcfe6b8e-1a14-4344-a03c-1f62c9b7baef. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.247148] env[61947]: DEBUG nova.network.neutron [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updating instance_info_cache with network_info: [{"id": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "address": "fa:16:3e:18:eb:2c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcfe6b8e-1a", "ovs_interfaceid": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.301967] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.315457] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.315883] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.316228] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Deleting the datastore file [datastore1] 7f308223-6c3c-4c62-8401-d7a44eeb69d2 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.316662] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cefda9db-c5e5-45c5-924b-683a2fef5b5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.326708] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for the task: (returnval){ [ 890.326708] env[61947]: value = "task-1224556" [ 890.326708] env[61947]: _type = "Task" [ 890.326708] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.340859] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.388160] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224553, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.399657] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224551, 'name': CreateVM_Task, 'duration_secs': 0.528716} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.399926] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.401126] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.401377] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.401845] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.404138] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4d7c667-a160-4e59-97ab-b9b6ad5f1c8b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.410988] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 890.410988] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5262d8b5-03aa-1876-403c-aa691e0797af" [ 890.410988] env[61947]: _type = "Task" [ 890.410988] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.421564] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5262d8b5-03aa-1876-403c-aa691e0797af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.540916] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224554, 'name': CreateVM_Task, 'duration_secs': 0.434508} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.541167] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.542275] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.542450] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.542770] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.543032] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa5d81f2-1eab-4ba4-bcd8-b88a853c9399 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.549032] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 890.549032] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524fbf13-966d-fdfe-2320-e1af6329ffbd" [ 890.549032] env[61947]: _type = "Task" [ 890.549032] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.558872] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524fbf13-966d-fdfe-2320-e1af6329ffbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.561236] env[61947]: DEBUG nova.scheduler.client.report [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.605484] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.606107] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.606107] env[61947]: DEBUG nova.network.neutron [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.750118] env[61947]: DEBUG oslo_concurrency.lockutils [req-6580a23a-0c55-4e25-b999-697269e26b45 req-993d297d-d7d9-4766-8031-b59cef9ee98f service nova] Releasing lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.840613] env[61947]: DEBUG oslo_vmware.api [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Task: {'id': task-1224556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199033} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.841531] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.842226] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.842371] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.842550] env[61947]: INFO nova.compute.manager [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 890.842790] env[61947]: DEBUG oslo.service.loopingcall [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.843232] env[61947]: DEBUG nova.compute.manager [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.843327] env[61947]: DEBUG nova.network.neutron [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.891089] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224553, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.926962] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5262d8b5-03aa-1876-403c-aa691e0797af, 'name': SearchDatastore_Task, 'duration_secs': 0.012073} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.927295] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.927642] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.927929] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.928128] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.928339] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.928723] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bd7e630-e8df-4550-a043-3a537b9bb261 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.939732] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.940030] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 890.941053] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44c69b5c-97a5-40e3-a82e-f67044adde62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.949690] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 890.949690] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52efe52e-c6c1-971f-1285-9ca2444d8d37" [ 890.949690] env[61947]: _type = "Task" [ 890.949690] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.965670] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52efe52e-c6c1-971f-1285-9ca2444d8d37, 'name': SearchDatastore_Task, 'duration_secs': 0.010595} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.966577] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c029cd-0aab-4f95-8173-93e0acfb6404 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.973947] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 890.973947] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5242a14d-5426-3e48-5deb-6f4272eee2a1" [ 890.973947] env[61947]: _type = "Task" [ 890.973947] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.982596] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5242a14d-5426-3e48-5deb-6f4272eee2a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.066247] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524fbf13-966d-fdfe-2320-e1af6329ffbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010775} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.066590] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.066832] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.067671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.067671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.067671] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.067914] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a41686e8-00e5-4d44-bde1-2266f3de6ef2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.075051] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.075051] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.076115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.063s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.077717] env[61947]: INFO nova.compute.claims [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.082037] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.082222] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.082998] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af722a47-c941-4cc1-aa83-4b51af3ae767 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.091217] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 891.091217] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524163ab-6a95-ba4f-4204-f419665b2396" [ 891.091217] env[61947]: _type = "Task" [ 891.091217] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.106139] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524163ab-6a95-ba4f-4204-f419665b2396, 'name': SearchDatastore_Task, 'duration_secs': 0.010416} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.110214] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff36046f-ce26-40aa-80c0-ef6f9c87208d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.119112] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 891.119112] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b4e22d-c24a-2c39-7dcd-295cbb774927" [ 891.119112] env[61947]: _type = "Task" [ 891.119112] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.126015] env[61947]: DEBUG nova.network.neutron [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updated VIF entry in instance network info cache for port 6832c702-f2d4-446f-a6e3-9225cf51f491. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.126545] env[61947]: DEBUG nova.network.neutron [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updating instance_info_cache with network_info: [{"id": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "address": "fa:16:3e:6e:4d:bc", "network": {"id": "b3f69232-c6a8-46da-a6c6-d9d3ec2ae81a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-522940052", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.179", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b59bc8-bc", "ovs_interfaceid": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6832c702-f2d4-446f-a6e3-9225cf51f491", "address": "fa:16:3e:0f:df:17", "network": {"id": "afbf9dc1-7cc0-4ce3-82f4-63d7f45a7950", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-514383215", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97113f46-d648-4613-b233-069acba18198", "external-id": "nsx-vlan-transportzone-480", "segmentation_id": 480, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832c702-f2", "ovs_interfaceid": "6832c702-f2d4-446f-a6e3-9225cf51f491", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.133229] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b4e22d-c24a-2c39-7dcd-295cbb774927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.328180] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.328413] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.328716] env[61947]: DEBUG nova.compute.manager [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Going to confirm migration 1 {{(pid=61947) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 891.388639] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224553, 'name': CloneVM_Task} progress is 95%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.480940] env[61947]: DEBUG nova.network.neutron [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.488996] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5242a14d-5426-3e48-5deb-6f4272eee2a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009244} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.489270] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.489579] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9cfa7c52-2250-49ce-a1c8-baad6644f824/9cfa7c52-2250-49ce-a1c8-baad6644f824.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 891.489843] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2a05a18-6aec-49c5-a70a-c678650c3376 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.499478] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "e0eebf59-c84e-4462-b280-d783a04525e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.499774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.500049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "e0eebf59-c84e-4462-b280-d783a04525e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.500277] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.500462] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.504348] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 891.504348] env[61947]: value = "task-1224558" [ 891.504348] env[61947]: _type = "Task" [ 891.504348] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.504796] env[61947]: INFO nova.compute.manager [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Terminating instance [ 891.509864] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "refresh_cache-e0eebf59-c84e-4462-b280-d783a04525e2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.510043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquired lock "refresh_cache-e0eebf59-c84e-4462-b280-d783a04525e2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.510231] env[61947]: DEBUG nova.network.neutron [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.517436] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.583608] env[61947]: DEBUG nova.compute.utils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.588927] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.588927] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.600186] env[61947]: DEBUG nova.network.neutron [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.634077] env[61947]: DEBUG oslo_concurrency.lockutils [req-7a0f147c-9fcf-4427-a832-c21d74f80740 req-46727396-e126-4562-8c87-a2cac29996f3 service nova] Releasing lock "refresh_cache-4f313bb5-5df4-4fb0-9736-95f4a519651d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.634687] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b4e22d-c24a-2c39-7dcd-295cbb774927, 'name': SearchDatastore_Task, 'duration_secs': 0.012482} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.635716] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.635716] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 4f313bb5-5df4-4fb0-9736-95f4a519651d/4f313bb5-5df4-4fb0-9736-95f4a519651d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 891.635716] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3795b0cd-034a-4965-a995-533788c1e840 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.645180] env[61947]: DEBUG nova.policy [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41d4a2d39ebc4ce0a0e76350f907e7df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58b83ea896bd4654aaa9e81f635a6a55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 891.648455] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 891.648455] env[61947]: value = "task-1224559" [ 891.648455] env[61947]: _type = "Task" [ 891.648455] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.658810] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.890047] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224553, 'name': CloneVM_Task, 'duration_secs': 1.7016} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.890253] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Created linked-clone VM from snapshot [ 891.891152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adee66d-4919-45c8-b9dd-185af55e95ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.901975] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploading image 1872e67e-0805-4b22-9fd4-036f704c08d5 {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 891.941197] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 891.941197] env[61947]: value = "vm-264674" [ 891.941197] env[61947]: _type = "VirtualMachine" [ 891.941197] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 891.941534] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8043ed7d-a6d9-40ad-af86-a57710d1fe09 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.944879] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.947219] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.947219] env[61947]: DEBUG nova.network.neutron [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.947219] env[61947]: DEBUG nova.objects.instance [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lazy-loading 'info_cache' on Instance uuid eede4f34-5ac4-475c-a74b-a98327f648fd {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.955958] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease: (returnval){ [ 891.955958] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af1be9-3aa9-7fc0-cb48-e4b5aad3a923" [ 891.955958] env[61947]: _type = "HttpNfcLease" [ 891.955958] env[61947]: } obtained for exporting VM: (result){ [ 891.955958] env[61947]: value = "vm-264674" [ 891.955958] env[61947]: _type = "VirtualMachine" [ 891.955958] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 891.956375] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the lease: (returnval){ [ 891.956375] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af1be9-3aa9-7fc0-cb48-e4b5aad3a923" [ 891.956375] env[61947]: _type = "HttpNfcLease" [ 891.956375] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 891.968165] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 891.968165] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af1be9-3aa9-7fc0-cb48-e4b5aad3a923" [ 891.968165] env[61947]: _type = "HttpNfcLease" [ 891.968165] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 891.984525] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.025045] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478236} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.025937] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Successfully created port: 9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.028408] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9cfa7c52-2250-49ce-a1c8-baad6644f824/9cfa7c52-2250-49ce-a1c8-baad6644f824.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 892.028732] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.029204] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da003012-4953-40d9-a1bb-180cbdd81b93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.041258] env[61947]: DEBUG nova.network.neutron [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.046023] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 892.046023] env[61947]: value = "task-1224561" [ 892.046023] env[61947]: _type = "Task" [ 892.046023] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.058514] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.093414] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.102432] env[61947]: INFO nova.compute.manager [-] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Took 1.26 seconds to deallocate network for instance. [ 892.126993] env[61947]: DEBUG nova.network.neutron [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.170550] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224559, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.421824] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464dafe0-2627-426b-a964-004401ee3aff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.432792] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067ca420-bdc3-414f-829d-c846cea266bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.473352] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061d0319-2a84-47c6-8ae2-0701ee1fb10d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.483263] env[61947]: DEBUG nova.compute.manager [req-9868c2d2-4757-4a13-9ba9-730991fbfee9 req-69e2b925-7f0e-4cdb-8727-ab19601afd5e service nova] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Received event network-vif-deleted-87e10fb3-d4e6-4b0e-aa08-af3ee4ace97b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.487200] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 892.487200] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af1be9-3aa9-7fc0-cb48-e4b5aad3a923" [ 892.487200] env[61947]: _type = "HttpNfcLease" [ 892.487200] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 892.488675] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13935a5-6353-4ad1-9688-fbdd1c40f2c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.494982] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 892.494982] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52af1be9-3aa9-7fc0-cb48-e4b5aad3a923" [ 892.494982] env[61947]: _type = "HttpNfcLease" [ 892.494982] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 892.500638] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaba7068-7421-4c6b-9ac1-db8b361bc545 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.520836] env[61947]: DEBUG nova.compute.provider_tree [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.520836] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 892.520836] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 892.523053] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e7ea61-ccad-44ae-bc86-5ec232e6dcd4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.609134] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c556da-85ac-4553-94ab-d7b076e56e6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.613387] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.613994] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149135} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.614617] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.615507] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fee0d1-8d4a-49be-8260-ebf1c2db4cbc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.621174] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 892.638950] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Releasing lock "refresh_cache-e0eebf59-c84e-4462-b280-d783a04525e2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.638950] env[61947]: DEBUG nova.compute.manager [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 892.638950] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 892.647362] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 9cfa7c52-2250-49ce-a1c8-baad6644f824/9cfa7c52-2250-49ce-a1c8-baad6644f824.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.648216] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8dad3a-e7ec-434e-b045-31afb26bbc69 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.652495] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-772cd84f-2a8e-4210-a1bb-f1022a6e5ce7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.676805] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224559, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.679904] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 892.679904] env[61947]: value = "task-1224562" [ 892.679904] env[61947]: _type = "Task" [ 892.679904] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.680403] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 892.680718] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ed1f82b-336a-4696-a162-5b9b062e1fa2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.692605] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224562, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.694574] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 892.694574] env[61947]: value = "task-1224563" [ 892.694574] env[61947]: _type = "Task" [ 892.694574] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.707045] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.718132] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-84f27bda-d77e-434e-8a3c-b13a00fa1c14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.023754] env[61947]: DEBUG nova.scheduler.client.report [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.111787] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.130014] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.133437] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.133822] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.134390] env[61947]: DEBUG nova.objects.instance [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'flavor' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.136131] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06a08c25-d0cd-4c18-9003-55b014d012b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.146781] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 893.146781] env[61947]: value = "task-1224564" [ 893.146781] env[61947]: _type = "Task" [ 893.146781] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.150305] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c091329bca1b2033b309a390b738b232',container_format='bare',created_at=2024-10-09T14:50:05Z,direct_url=,disk_format='vmdk',id=13296250-6feb-41b0-b80f-26c6dd557dde,min_disk=1,min_ram=0,name='tempest-test-snap-560750883',owner='58b83ea896bd4654aaa9e81f635a6a55',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-09T14:50:21Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.150874] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.151118] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.151341] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.152305] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.152468] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.152690] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.152861] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.153051] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.153226] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.153409] env[61947]: DEBUG nova.virt.hardware [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.154352] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8ab337-e30d-4f0d-abe9-118843b137ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.178443] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224564, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.180141] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74a1b36-a0f6-4f5a-a1bd-5533d80208fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.189862] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224559, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.212278] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224562, 'name': ReconfigVM_Task, 'duration_secs': 0.326144} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.213595] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 9cfa7c52-2250-49ce-a1c8-baad6644f824/9cfa7c52-2250-49ce-a1c8-baad6644f824.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.214384] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36f1b558-9427-40e3-b6d1-40bdddb5a3b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.220483] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224563, 'name': PowerOffVM_Task, 'duration_secs': 0.203809} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.224304] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.224304] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 893.224304] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4cc43ae-872f-4ad2-a55a-2569087fbeef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.232082] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 893.232082] env[61947]: value = "task-1224565" [ 893.232082] env[61947]: _type = "Task" [ 893.232082] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.242216] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224565, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.260161] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 893.260293] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 893.260474] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Deleting the datastore file [datastore1] e0eebf59-c84e-4462-b280-d783a04525e2 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.260749] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35e04af9-94b3-4245-8aca-1237d32202a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.265518] env[61947]: DEBUG nova.network.neutron [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [{"id": "5e37697a-486f-4961-aed8-8ccdbf59992b", "address": "fa:16:3e:35:27:73", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e37697a-48", "ovs_interfaceid": "5e37697a-486f-4961-aed8-8ccdbf59992b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.268905] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for the task: (returnval){ [ 893.268905] env[61947]: value = "task-1224567" [ 893.268905] env[61947]: _type = "Task" [ 893.268905] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.278906] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224567, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.532507] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.533413] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.538028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.873s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.538028] env[61947]: DEBUG nova.objects.instance [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lazy-loading 'resources' on Instance uuid fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.641605] env[61947]: DEBUG nova.objects.instance [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'pci_requests' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.670945] env[61947]: DEBUG oslo_vmware.api [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224564, 'name': PowerOnVM_Task, 'duration_secs': 0.511079} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.676647] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 893.679297] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a25014af-12a6-44b9-9b87-aa1468c3af5f tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance '72818e40-624b-4c04-888b-bb622f7f96d7' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 893.693503] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224559, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.700102} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.694464] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 4f313bb5-5df4-4fb0-9736-95f4a519651d/4f313bb5-5df4-4fb0-9736-95f4a519651d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.694464] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.694464] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07e11efc-60c8-4053-9430-0c5a953375fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.704024] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 893.704024] env[61947]: value = "task-1224569" [ 893.704024] env[61947]: _type = "Task" [ 893.704024] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.715385] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224569, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.743918] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224565, 'name': Rename_Task, 'duration_secs': 0.200298} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.745792] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.746226] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-337a4ca4-3a0f-411c-92ae-3a1bc24cbc9d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.770149] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-eede4f34-5ac4-475c-a74b-a98327f648fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.770591] env[61947]: DEBUG nova.objects.instance [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lazy-loading 'migration_context' on Instance uuid eede4f34-5ac4-475c-a74b-a98327f648fd {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.778310] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 893.778310] env[61947]: value = "task-1224570" [ 893.778310] env[61947]: _type = "Task" [ 893.778310] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.789055] env[61947]: DEBUG oslo_vmware.api [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Task: {'id': task-1224567, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109005} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.789055] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.789430] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 893.791438] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 893.791438] env[61947]: INFO nova.compute.manager [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 893.791438] env[61947]: DEBUG oslo.service.loopingcall [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.791438] env[61947]: DEBUG nova.compute.manager [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.791438] env[61947]: DEBUG nova.network.neutron [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 893.793527] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Successfully updated port: 9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.800165] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.823114] env[61947]: DEBUG nova.network.neutron [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.042023] env[61947]: DEBUG nova.objects.instance [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lazy-loading 'numa_topology' on Instance uuid fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.044175] env[61947]: DEBUG nova.compute.utils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.046118] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.046290] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.117941] env[61947]: DEBUG nova.policy [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a08722af783a4ed5b6f6a48a9e2a7c24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86f19f3f002c4686a26d843a4c819d89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 894.144397] env[61947]: DEBUG nova.objects.base [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Object Instance<04e5cc55-71d9-4d3d-95c1-fb1401ab74f8> lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.144653] env[61947]: DEBUG nova.network.neutron [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.215035] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224569, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076298} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.216937] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.217827] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e525d26-9eb6-4260-8eea-4d8963e6034c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.252230] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 4f313bb5-5df4-4fb0-9736-95f4a519651d/4f313bb5-5df4-4fb0-9736-95f4a519651d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.254080] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6170ef3b-4f96-4229-b6e2-93446800b82b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.120s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.255445] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-937ba299-320d-489a-b205-dba59e244d8f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.273398] env[61947]: DEBUG nova.objects.base [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.275152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2a79c3-8bf4-4075-bdfe-f087fbd22d2f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.284473] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 894.284473] env[61947]: value = "task-1224571" [ 894.284473] env[61947]: _type = "Task" [ 894.284473] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.307974] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.308292] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.308585] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.319109] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c343241e-9607-4bcc-9394-b8ef58749160 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.322118] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224571, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.322397] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.327376] env[61947]: DEBUG nova.network.neutron [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.331524] env[61947]: DEBUG oslo_vmware.api [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 894.331524] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520ec9ca-971f-8591-3417-99b3960affa6" [ 894.331524] env[61947]: _type = "Task" [ 894.331524] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.345265] env[61947]: DEBUG oslo_vmware.api [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520ec9ca-971f-8591-3417-99b3960affa6, 'name': SearchDatastore_Task, 'duration_secs': 0.007433} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.345681] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.547596] env[61947]: DEBUG nova.objects.base [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 894.552129] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.555886] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Successfully created port: ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.794951] env[61947]: DEBUG oslo_vmware.api [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224570, 'name': PowerOnVM_Task, 'duration_secs': 0.855014} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.795654] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 894.795993] env[61947]: INFO nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Took 8.08 seconds to spawn the instance on the hypervisor. [ 894.796350] env[61947]: DEBUG nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.798437] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7241cf-4340-4b5c-a3d9-5601fdf1121d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.804956] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224571, 'name': ReconfigVM_Task, 'duration_secs': 0.289587} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.808666] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 4f313bb5-5df4-4fb0-9736-95f4a519651d/4f313bb5-5df4-4fb0-9736-95f4a519651d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.809947] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eaa753b-6fb1-4122-860a-c1271b4190bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.834152] env[61947]: INFO nova.compute.manager [-] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Took 1.04 seconds to deallocate network for instance. [ 894.835822] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 894.835822] env[61947]: value = "task-1224572" [ 894.835822] env[61947]: _type = "Task" [ 894.835822] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.848381] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903321f2-382e-403d-ab69-26a56273f42e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.857449] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224572, 'name': Rename_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.861603] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.868749] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4be1e80-4862-4dfa-8d0e-210d652ec19b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.908468] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5318fade-f1ab-4d89-a44a-74ff4ade82e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.922031] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592c6037-d052-4849-a165-fb7c0ba53b27 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.943672] env[61947]: DEBUG nova.compute.provider_tree [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.986311] env[61947]: DEBUG nova.compute.manager [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Received event network-vif-plugged-9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.986543] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Acquiring lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.986769] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.986945] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.987164] env[61947]: DEBUG nova.compute.manager [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] No waiting events found dispatching network-vif-plugged-9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.987365] env[61947]: WARNING nova.compute.manager [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Received unexpected event network-vif-plugged-9de62525-ad88-46d7-9cab-b679b8515e82 for instance with vm_state building and task_state spawning. [ 894.987536] env[61947]: DEBUG nova.compute.manager [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Received event network-changed-9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.987691] env[61947]: DEBUG nova.compute.manager [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Refreshing instance network info cache due to event network-changed-9de62525-ad88-46d7-9cab-b679b8515e82. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 894.987858] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Acquiring lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.059342] env[61947]: DEBUG nova.network.neutron [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Updating instance_info_cache with network_info: [{"id": "9de62525-ad88-46d7-9cab-b679b8515e82", "address": "fa:16:3e:00:23:f2", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de62525-ad", "ovs_interfaceid": "9de62525-ad88-46d7-9cab-b679b8515e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.337071] env[61947]: INFO nova.compute.manager [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Took 27.24 seconds to build instance. [ 895.349233] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224572, 'name': Rename_Task, 'duration_secs': 0.201738} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.350207] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.350511] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae4f74c8-c7b4-4618-af91-9b2c3bc1a80d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.353305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.360322] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 895.360322] env[61947]: value = "task-1224573" [ 895.360322] env[61947]: _type = "Task" [ 895.360322] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.369688] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224573, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.450876] env[61947]: DEBUG nova.scheduler.client.report [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.565311] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.565669] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Instance network_info: |[{"id": "9de62525-ad88-46d7-9cab-b679b8515e82", "address": "fa:16:3e:00:23:f2", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de62525-ad", "ovs_interfaceid": "9de62525-ad88-46d7-9cab-b679b8515e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.566784] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.569851] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Acquired lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.570449] env[61947]: DEBUG nova.network.neutron [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Refreshing network info cache for port 9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.571686] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:23:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9de62525-ad88-46d7-9cab-b679b8515e82', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.580518] env[61947]: DEBUG oslo.service.loopingcall [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.583822] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.584394] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46722620-e817-42f7-818a-e155a6cbe227 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.609387] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.609763] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.609865] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.609985] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.610153] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.610730] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.610730] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.610730] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.610926] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.611022] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.611231] env[61947]: DEBUG nova.virt.hardware [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.612074] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdd1579-4da7-4c19-98c1-ec8f7aca4ac4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.618176] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.618176] env[61947]: value = "task-1224574" [ 895.618176] env[61947]: _type = "Task" [ 895.618176] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.625174] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214df048-734f-41b6-a42f-85f27ad11e35 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.633708] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224574, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.840878] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27c461d5-7034-48c1-8a98-428b2d4aa781 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.749s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.875232] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224573, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.956597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.419s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.960491] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.380s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.963125] env[61947]: INFO nova.compute.claims [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.971041] env[61947]: DEBUG nova.network.neutron [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Updated VIF entry in instance network info cache for port 9de62525-ad88-46d7-9cab-b679b8515e82. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.972183] env[61947]: DEBUG nova.network.neutron [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Updating instance_info_cache with network_info: [{"id": "9de62525-ad88-46d7-9cab-b679b8515e82", "address": "fa:16:3e:00:23:f2", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9de62525-ad", "ovs_interfaceid": "9de62525-ad88-46d7-9cab-b679b8515e82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.136556] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224574, 'name': CreateVM_Task, 'duration_secs': 0.413973} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.137430] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.139180] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.139727] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.140905] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.142576] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def4b55a-6596-48c9-bbc8-a57c1958ad7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.149584] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 896.149584] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521f345a-683d-db15-3e10-22d134cdb042" [ 896.149584] env[61947]: _type = "Task" [ 896.149584] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.171315] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.171787] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Processing image 13296250-6feb-41b0-b80f-26c6dd557dde {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.172255] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.172597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.172990] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.173489] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2a55c19-b271-46db-91b5-5facea1c6f88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.198261] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.198876] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.199982] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e405eef8-f64b-41eb-b128-45aaeb919a63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.207115] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 896.207115] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e70d35-5c2d-f45d-a601-cef69d7f50b3" [ 896.207115] env[61947]: _type = "Task" [ 896.207115] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.216596] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e70d35-5c2d-f45d-a601-cef69d7f50b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.267356] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Successfully updated port: ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.373197] env[61947]: DEBUG oslo_vmware.api [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224573, 'name': PowerOnVM_Task, 'duration_secs': 0.522373} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.373551] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.373789] env[61947]: INFO nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Took 12.09 seconds to spawn the instance on the hypervisor. [ 896.373988] env[61947]: DEBUG nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.374891] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43194d9f-3f84-444a-bfba-b8b5814c57ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.479736] env[61947]: DEBUG oslo_concurrency.lockutils [req-93a205ba-7ab0-45be-999a-5427d35164eb req-87bb5eb5-2dd9-4566-b64a-d86968a2c25e service nova] Releasing lock "refresh_cache-f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.479736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e6dfa1d4-a462-424c-a622-f316108358ea tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.563s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.479736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.482s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.479736] env[61947]: INFO nova.compute.manager [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Unshelving [ 896.638119] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.638442] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.638718] env[61947]: INFO nova.compute.manager [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Shelving [ 896.642035] env[61947]: DEBUG nova.compute.manager [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-vif-plugged-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.642277] env[61947]: DEBUG oslo_concurrency.lockutils [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] Acquiring lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.642499] env[61947]: DEBUG oslo_concurrency.lockutils [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.642755] env[61947]: DEBUG oslo_concurrency.lockutils [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.642958] env[61947]: DEBUG nova.compute.manager [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] No waiting events found dispatching network-vif-plugged-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.643146] env[61947]: WARNING nova.compute.manager [req-7daa4555-f684-4977-88e6-aa77976023a3 req-bb1d9fe4-03d4-4f53-b36b-c9f25847ace0 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received unexpected event network-vif-plugged-ab2d3df7-884a-42c4-83db-ff3f86592667 for instance with vm_state building and task_state spawning. [ 896.684761] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.685178] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.685649] env[61947]: DEBUG nova.objects.instance [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'flavor' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.719036] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 896.719211] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Fetch image to [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899/OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 896.719334] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Downloading stream optimized image 13296250-6feb-41b0-b80f-26c6dd557dde to [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899/OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899.vmdk on the data store datastore1 as vApp {{(pid=61947) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 896.719512] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Downloading image file data 13296250-6feb-41b0-b80f-26c6dd557dde to the ESX as VM named 'OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899' {{(pid=61947) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 896.770134] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.770134] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.770408] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.799022] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 896.799022] env[61947]: value = "resgroup-9" [ 896.799022] env[61947]: _type = "ResourcePool" [ 896.799022] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 896.799334] env[61947]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a90e1a1a-01ff-4880-bb1b-1cd4b60feec4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.831269] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease: (returnval){ [ 896.831269] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 896.831269] env[61947]: _type = "HttpNfcLease" [ 896.831269] env[61947]: } obtained for vApp import into resource pool (val){ [ 896.831269] env[61947]: value = "resgroup-9" [ 896.831269] env[61947]: _type = "ResourcePool" [ 896.831269] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 896.831724] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the lease: (returnval){ [ 896.831724] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 896.831724] env[61947]: _type = "HttpNfcLease" [ 896.831724] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 896.840136] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.840136] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 896.840136] env[61947]: _type = "HttpNfcLease" [ 896.840136] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.894888] env[61947]: INFO nova.compute.manager [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Took 30.03 seconds to build instance. [ 897.150773] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.151392] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee474e61-f202-40e7-bd99-84de857bcee3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.160189] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 897.160189] env[61947]: value = "task-1224576" [ 897.160189] env[61947]: _type = "Task" [ 897.160189] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.161029] env[61947]: DEBUG nova.network.neutron [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Port c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 897.161273] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.161450] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.161617] env[61947]: DEBUG nova.network.neutron [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 897.179059] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.226746] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c61790-22b1-4134-9f9d-fb8ba53d4ded {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.240966] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c757a12-9b48-49de-87b0-8fba9c7aafaa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.272614] env[61947]: DEBUG nova.objects.instance [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'pci_requests' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.274241] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241fac6f-8a86-4fce-8957-be4133c285cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.285752] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b584c9b1-c27c-4c5f-a85f-6a9515ef9c07 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.301949] env[61947]: DEBUG nova.compute.provider_tree [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.311563] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.340454] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.340454] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 897.340454] env[61947]: _type = "HttpNfcLease" [ 897.340454] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 897.397290] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3538a7a1-7902-40fe-8354-7aabda628f94 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.546s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.446252] env[61947]: DEBUG nova.network.neutron [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.505620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.675502] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224576, 'name': PowerOffVM_Task, 'duration_secs': 0.278017} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.676354] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.677224] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3db2285-bc30-4ec3-b7c9-a96b88c1f511 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.700322] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378515d7-f682-4710-b64d-49cc6220aeb0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.780310] env[61947]: DEBUG nova.objects.base [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Object Instance<04e5cc55-71d9-4d3d-95c1-fb1401ab74f8> lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 897.780500] env[61947]: DEBUG nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.804778] env[61947]: DEBUG nova.scheduler.client.report [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.841639] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.841639] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 897.841639] env[61947]: _type = "HttpNfcLease" [ 897.841639] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 897.850106] env[61947]: DEBUG nova.policy [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 897.878445] env[61947]: DEBUG nova.network.neutron [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.950096] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.950484] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Instance network_info: |[{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.951013] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:19:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e0c77754-4085-434b-a3e8-d61be099ac67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab2d3df7-884a-42c4-83db-ff3f86592667', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.959605] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Creating folder: Project (86f19f3f002c4686a26d843a4c819d89). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 897.961167] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de9d05f9-5bc1-480e-b118-696b1c6cba61 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.974969] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Created folder: Project (86f19f3f002c4686a26d843a4c819d89) in parent group-v264556. [ 897.975362] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Creating folder: Instances. Parent ref: group-v264678. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 897.975706] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be532209-d86d-4d8a-94e4-2f17ff6e2172 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.986235] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Created folder: Instances in parent group-v264678. [ 897.986495] env[61947]: DEBUG oslo.service.loopingcall [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.986703] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.986932] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-007e8f9f-1a30-4366-bbf8-460253422fa5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.008369] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.008369] env[61947]: value = "task-1224579" [ 898.008369] env[61947]: _type = "Task" [ 898.008369] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.018368] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224579, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.215132] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 898.215520] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2f420af9-1f1a-4549-8c6d-5b04026d7c70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.227043] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 898.227043] env[61947]: value = "task-1224580" [ 898.227043] env[61947]: _type = "Task" [ 898.227043] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.242848] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224580, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.256250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "91205e54-6bcb-403c-8308-0f0692fe4fba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.256507] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.295713] env[61947]: DEBUG nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Successfully created port: 4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.312260] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.312861] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.316041] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.014s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.316196] env[61947]: DEBUG nova.objects.instance [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61947) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 898.343203] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 898.343203] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 898.343203] env[61947]: _type = "HttpNfcLease" [ 898.343203] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 898.382989] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.520539] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224579, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.530020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.530020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.737732] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224580, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.760717] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.820618] env[61947]: DEBUG nova.compute.utils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.825027] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.825141] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.844346] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 898.844346] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 898.844346] env[61947]: _type = "HttpNfcLease" [ 898.844346] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 898.878147] env[61947]: DEBUG nova.policy [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15d31bf1d0bb46fcabbafbe722265877', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc5cf1feb86a41fcbfb0b1b96d90998e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 898.885781] env[61947]: DEBUG nova.compute.manager [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61947) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 898.885781] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.019291] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224579, 'name': CreateVM_Task, 'duration_secs': 0.866537} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.019774] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 899.020223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.020394] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.020732] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.021016] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3a2976c-48e0-4fd0-8bfa-fbbf5169ac9a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.027127] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 899.027127] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d3df6c-8162-e62e-1270-024cd8b336aa" [ 899.027127] env[61947]: _type = "Task" [ 899.027127] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.032841] env[61947]: DEBUG nova.compute.utils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.037690] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d3df6c-8162-e62e-1270-024cd8b336aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.197756] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Successfully created port: 0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.238680] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224580, 'name': CreateSnapshot_Task, 'duration_secs': 0.744161} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.238986] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.239762] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99ef7b4-a659-4288-9688-8239f5ef7041 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.285281] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.326030] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.333482] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7169d6d2-4628-4d46-9452-188bebbb4eab tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.333482] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.719s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.333482] env[61947]: DEBUG nova.objects.instance [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lazy-loading 'resources' on Instance uuid 7f308223-6c3c-4c62-8401-d7a44eeb69d2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.344823] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.344823] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 899.344823] env[61947]: _type = "HttpNfcLease" [ 899.344823] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 899.345799] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 899.345799] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a303c0-c736-911e-4b8f-43215564ead5" [ 899.345799] env[61947]: _type = "HttpNfcLease" [ 899.345799] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 899.347161] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599715bf-c0ee-4999-a392-2992e80e7286 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.358244] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 899.358831] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 899.427412] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-71571144-aede-4580-b770-f11d0d7fedba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.538960] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.539619] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d3df6c-8162-e62e-1270-024cd8b336aa, 'name': SearchDatastore_Task, 'duration_secs': 0.01345} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.539981] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.540266] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.540553] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.540730] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.540947] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.542960] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d087008-0610-492b-bd10-160195437285 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.552406] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.552600] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 899.553665] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-000dd143-d663-433d-ad08-7f2c2376974e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.559716] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 899.559716] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520343d7-1f76-5784-6d2c-1bcbaf40bbad" [ 899.559716] env[61947]: _type = "Task" [ 899.559716] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.572025] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520343d7-1f76-5784-6d2c-1bcbaf40bbad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.764364] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.764364] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-82300a4c-4d4e-496f-8950-578289364c34 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.777024] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 899.777024] env[61947]: value = "task-1224581" [ 899.777024] env[61947]: _type = "Task" [ 899.777024] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.787525] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224581, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.917690] env[61947]: DEBUG nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Successfully updated port: 4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.075706] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520343d7-1f76-5784-6d2c-1bcbaf40bbad, 'name': SearchDatastore_Task, 'duration_secs': 0.015269} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.081191] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7088458-cb30-4bed-b815-3ffeb342be7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.090360] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 900.090360] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52588352-56f8-94f6-1d6f-a6ba253e4b8a" [ 900.090360] env[61947]: _type = "Task" [ 900.090360] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.107838] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52588352-56f8-94f6-1d6f-a6ba253e4b8a, 'name': SearchDatastore_Task, 'duration_secs': 0.01135} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.111238] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.111831] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/055a0b7f-1703-447c-b9d4-0a8efc4f5dbb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 900.114915] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20df8598-d108-4ca5-907f-fc6f0137ef37 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.130644] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 900.130644] env[61947]: value = "task-1224582" [ 900.130644] env[61947]: _type = "Task" [ 900.130644] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.147374] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.153421] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d21fb1-3771-410c-95ab-96f9a1f7ad0d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.167015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999ae509-5b5d-4d43-9076-5b1797224b33 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.205826] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329db02f-3efc-4d84-9dde-c2eb01278467 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.215449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6d1d06-0d6f-450a-b0b6-9ab7ddf446d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.232841] env[61947]: DEBUG nova.compute.provider_tree [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.290187] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224581, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.341933] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.374534] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.375388] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.375388] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.375570] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.375626] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.375860] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.376206] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.376469] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.376712] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.376916] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.377115] env[61947]: DEBUG nova.virt.hardware [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.378415] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b67ec29-5643-4659-b51b-534fdcce102c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.383900] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 900.384147] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 900.385057] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17610f83-cbb8-48eb-afe3-2e220e2e26b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.398303] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fbf430f-147d-4acd-b895-905716cd418d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.403079] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 900.403378] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 900.404136] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6712ce6c-2746-4df4-8ab5-e8f26c5496a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.421785] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.422038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.422581] env[61947]: DEBUG nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.626061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.626411] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.626714] env[61947]: INFO nova.compute.manager [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Attaching volume 51d58e15-9bd1-43c0-a542-8210c2873e98 to /dev/sdb [ 900.646644] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224582, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.670899] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a744fea5-0195-44af-8f25-a1449a128bd7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.679783] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50969680-e39c-440f-824e-3166b467d23b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.697367] env[61947]: DEBUG nova.virt.block_device [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updating existing volume attachment record: fb98e257-3ae8-46a0-9d67-084b04b0e295 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 900.739796] env[61947]: DEBUG nova.scheduler.client.report [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.750502] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Successfully updated port: 0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.791584] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224581, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.839684] env[61947]: DEBUG oslo_vmware.rw_handles [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524d98cb-1eb0-0df1-d473-66c8cc70d2c1/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 900.840156] env[61947]: INFO nova.virt.vmwareapi.images [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Downloaded image file data 13296250-6feb-41b0-b80f-26c6dd557dde [ 900.841638] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bc7634-543a-4b07-835c-84a617f5775e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.869204] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f3af54f-356f-4275-8d22-fef6d5f5d190 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.901455] env[61947]: INFO nova.virt.vmwareapi.images [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] The imported VM was unregistered [ 900.903326] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 900.903493] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating directory with path [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.903771] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e68c89a-142b-4691-b095-6fc0e84ecb50 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.926836] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created directory with path [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.926988] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899/OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899.vmdk to [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk. {{(pid=61947) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 900.927270] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a414c851-4183-4928-b7cc-5cbc0aeccf0a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.938944] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 900.938944] env[61947]: value = "task-1224586" [ 900.938944] env[61947]: _type = "Task" [ 900.938944] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.948777] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.964958] env[61947]: DEBUG nova.compute.manager [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.965201] env[61947]: DEBUG nova.compute.manager [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing instance network info cache due to event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.965414] env[61947]: DEBUG oslo_concurrency.lockutils [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.965641] env[61947]: DEBUG oslo_concurrency.lockutils [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.965856] env[61947]: DEBUG nova.network.neutron [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.967740] env[61947]: WARNING nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] 720a7469-85cf-4e25-82f3-b574730aae1f already exists in list: networks containing: ['720a7469-85cf-4e25-82f3-b574730aae1f']. ignoring it [ 900.985602] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 900.987036] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4b933c-99d8-42ff-985d-35b57e7bda1c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.994210] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 900.994406] env[61947]: ERROR oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk due to incomplete transfer. [ 900.994950] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a9d7de0-fc9e-4fcc-b400-f5fed8c43bc8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.003407] env[61947]: DEBUG oslo_vmware.rw_handles [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523110b0-e6ff-8402-991e-b6fee667edb4/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 901.003604] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Uploaded image 1872e67e-0805-4b22-9fd4-036f704c08d5 to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 901.005977] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 901.006266] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ae2c3b07-2123-4cc5-ac35-c62baa80166e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.017920] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 901.017920] env[61947]: value = "task-1224588" [ 901.017920] env[61947]: _type = "Task" [ 901.017920] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.027020] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224588, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.145090] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553415} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.145090] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/055a0b7f-1703-447c-b9d4-0a8efc4f5dbb.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 901.145090] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.145090] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c3a0df7e-8bba-4677-8be6-e0de9727aab1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.154352] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 901.154352] env[61947]: value = "task-1224589" [ 901.154352] env[61947]: _type = "Task" [ 901.154352] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.164205] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.252022] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.919s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.254039] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.254146] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.254323] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.255758] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.910s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.296898] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224581, 'name': CloneVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.309438] env[61947]: INFO nova.scheduler.client.report [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Deleted allocations for instance 7f308223-6c3c-4c62-8401-d7a44eeb69d2 [ 901.385869] env[61947]: DEBUG nova.network.neutron [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "address": "fa:16:3e:2f:9a:dc", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1cb946-73", "ovs_interfaceid": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.452712] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.532656] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224588, 'name': Destroy_Task, 'duration_secs': 0.345825} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.532970] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroyed the VM [ 901.533258] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 901.533555] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9739f1ad-79f3-4812-8412-630f4b4ebf98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.548108] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 901.548108] env[61947]: value = "task-1224590" [ 901.548108] env[61947]: _type = "Task" [ 901.548108] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.562844] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224590, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.668624] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075074} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.669061] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.669951] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f432de-19b0-45df-a6c7-5469f5272cd7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.699084] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/055a0b7f-1703-447c-b9d4-0a8efc4f5dbb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.699721] env[61947]: DEBUG nova.network.neutron [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updated VIF entry in instance network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.700086] env[61947]: DEBUG nova.network.neutron [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.703763] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6456b9fe-4a53-402c-bbab-5be92c1a7212 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.729898] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 901.729898] env[61947]: value = "task-1224591" [ 901.729898] env[61947]: _type = "Task" [ 901.729898] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.742587] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224591, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.795275] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224581, 'name': CloneVM_Task, 'duration_secs': 1.591179} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.795599] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Created linked-clone VM from snapshot [ 901.796649] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6b3de9-ea2f-4fee-8374-27dba62438c6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.803785] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.809475] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Uploading image e4fc5b9a-582d-4cb5-b647-324f95b10a34 {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 901.828837] env[61947]: DEBUG oslo_concurrency.lockutils [None req-648c4214-b216-4938-8391-b3358873f185 tempest-ServerGroupTestJSON-1769664629 tempest-ServerGroupTestJSON-1769664629-project-member] Lock "7f308223-6c3c-4c62-8401-d7a44eeb69d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.142s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.839785] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.839785] env[61947]: value = "vm-264682" [ 901.839785] env[61947]: _type = "VirtualMachine" [ 901.839785] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.840202] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4255c660-a911-4bd0-8617-466e2b02d730 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.847405] env[61947]: DEBUG nova.compute.manager [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-plugged-4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.847688] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.848018] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.848264] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.848453] env[61947]: DEBUG nova.compute.manager [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] No waiting events found dispatching network-vif-plugged-4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.848682] env[61947]: WARNING nova.compute.manager [req-f5f1847f-f331-4ae3-93c1-b4a80c1afde3 req-718d61da-5dfa-473f-816d-0b46f613ce20 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received unexpected event network-vif-plugged-4e1cb946-7339-4abc-a85d-28eb52415eb9 for instance with vm_state active and task_state None. [ 901.859752] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lease: (returnval){ [ 901.859752] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ee971f-9078-e59c-79f7-48a572333e8b" [ 901.859752] env[61947]: _type = "HttpNfcLease" [ 901.859752] env[61947]: } obtained for exporting VM: (result){ [ 901.859752] env[61947]: value = "vm-264682" [ 901.859752] env[61947]: _type = "VirtualMachine" [ 901.859752] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.860097] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the lease: (returnval){ [ 901.860097] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ee971f-9078-e59c-79f7-48a572333e8b" [ 901.860097] env[61947]: _type = "HttpNfcLease" [ 901.860097] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.875952] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.875952] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ee971f-9078-e59c-79f7-48a572333e8b" [ 901.875952] env[61947]: _type = "HttpNfcLease" [ 901.875952] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.889513] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.891047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.891047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.894419] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b897045-cda2-4908-8d12-d7f2fc44eaeb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.917905] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.918235] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.918415] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.918761] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.918996] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.919185] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.919408] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.919582] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.919869] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.920159] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.920356] env[61947]: DEBUG nova.virt.hardware [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.928032] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfiguring VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 901.933109] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff94b5bd-364e-4027-869a-4131795383dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.975213] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.975213] env[61947]: DEBUG oslo_vmware.api [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 901.975213] env[61947]: value = "task-1224593" [ 901.975213] env[61947]: _type = "Task" [ 901.975213] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.989351] env[61947]: DEBUG oslo_vmware.api [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224593, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.995428] env[61947]: DEBUG nova.network.neutron [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.068191] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224590, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.073089] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9565ff-d068-4778-9027-8da6479c47f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.083112] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8798beb-2769-4a60-9fd5-bba63da82343 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.119733] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65d1fd1-3562-401d-9aee-95a2768f4d5c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.132078] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d231427d-78e0-4e84-84a3-114c5c76eb19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.153165] env[61947]: DEBUG nova.compute.provider_tree [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.220041] env[61947]: DEBUG oslo_concurrency.lockutils [req-27c0b9a4-fdfe-47d4-bf5e-d7bd4f1f643c req-bdef9a07-fd42-404e-a354-7a9ca9137960 service nova] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.246586] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.372507] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.372507] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ee971f-9078-e59c-79f7-48a572333e8b" [ 902.372507] env[61947]: _type = "HttpNfcLease" [ 902.372507] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.373108] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.373108] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ee971f-9078-e59c-79f7-48a572333e8b" [ 902.373108] env[61947]: _type = "HttpNfcLease" [ 902.373108] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.374068] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30889ed-5ce9-47f8-b70d-05e2b83798bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.387508] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.387899] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.456843] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.457049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.457267] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.457490] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.457733] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.461172] env[61947]: INFO nova.compute.manager [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Terminating instance [ 902.463530] env[61947]: DEBUG nova.compute.manager [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.463802] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.464758] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1fdbd4-ca55-4e5d-b943-f9466e9d7d7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.471925] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.482855] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 902.483150] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01c0f488-8ee3-44e6-99e9-2a1120379172 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.488568] env[61947]: DEBUG oslo_vmware.api [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.500367] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 902.500367] env[61947]: value = "task-1224594" [ 902.500367] env[61947]: _type = "Task" [ 902.500367] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.504701] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.504913] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Instance network_info: |[{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.505428] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:b6:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b7626fb-f420-4ad3-bea5-09a0daafed10', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.514477] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Creating folder: Project (dc5cf1feb86a41fcbfb0b1b96d90998e). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 902.515346] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ef63994-a578-42fb-b901-2336ae384999 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.529122] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.531069] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Created folder: Project (dc5cf1feb86a41fcbfb0b1b96d90998e) in parent group-v264556. [ 902.531349] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Creating folder: Instances. Parent ref: group-v264685. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 902.531702] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dc24617-876a-4c1d-8097-6de87010f67d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.536868] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4edc0b56-0163-40d6-8b2a-2758cafc23a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.554595] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Created folder: Instances in parent group-v264685. [ 902.554961] env[61947]: DEBUG oslo.service.loopingcall [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.558042] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.558963] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e078c8d-c145-42d2-9034-9d78aab144f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.584312] env[61947]: DEBUG oslo_vmware.api [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224590, 'name': RemoveSnapshot_Task, 'duration_secs': 0.6102} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.585440] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 902.585716] env[61947]: INFO nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 14.84 seconds to snapshot the instance on the hypervisor. [ 902.597232] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.597232] env[61947]: value = "task-1224597" [ 902.597232] env[61947]: _type = "Task" [ 902.597232] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.609586] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224597, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.657056] env[61947]: DEBUG nova.scheduler.client.report [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.746198] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.966191] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.991523] env[61947]: DEBUG oslo_vmware.api [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224593, 'name': ReconfigVM_Task, 'duration_secs': 0.707531} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.992085] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.992325] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfigured VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 903.013133] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.111687] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224597, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.138310] env[61947]: DEBUG nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Found 3 images (rotation: 2) {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 903.138824] env[61947]: DEBUG nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Rotating out 1 backups {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 903.139299] env[61947]: DEBUG nova.compute.manager [None req-0bcfacc2-d726-4bcb-be2e-850065cf9c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleting image 1cb3a519-7891-4eeb-9474-821506e882fb {{(pid=61947) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 903.249684] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224591, 'name': ReconfigVM_Task, 'duration_secs': 1.462175} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.249684] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/055a0b7f-1703-447c-b9d4-0a8efc4f5dbb.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.250354] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb82e738-952e-40ac-a942-af84bbbcd7aa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.261238] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 903.261238] env[61947]: value = "task-1224599" [ 903.261238] env[61947]: _type = "Task" [ 903.261238] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.277920] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224599, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.401449] env[61947]: DEBUG nova.compute.manager [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Received event network-vif-plugged-0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.402055] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.402305] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.402513] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.402880] env[61947]: DEBUG nova.compute.manager [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] No waiting events found dispatching network-vif-plugged-0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.403159] env[61947]: WARNING nova.compute.manager [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Received unexpected event network-vif-plugged-0b7626fb-f420-4ad3-bea5-09a0daafed10 for instance with vm_state building and task_state spawning. [ 903.403393] env[61947]: DEBUG nova.compute.manager [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Received event network-changed-0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.403607] env[61947]: DEBUG nova.compute.manager [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Refreshing instance network info cache due to event network-changed-0b7626fb-f420-4ad3-bea5-09a0daafed10. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 903.403885] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.404339] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.404586] env[61947]: DEBUG nova.network.neutron [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Refreshing network info cache for port 0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.467339] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.499716] env[61947]: DEBUG oslo_concurrency.lockutils [None req-936b74e3-1d6a-4d91-9f4f-c6324df7c43f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.813s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.515696] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.612289] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224597, 'name': CreateVM_Task, 'duration_secs': 0.526705} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.612561] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.613567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.613971] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.614568] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.615431] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43ba52a4-bc8e-491f-a14c-2256ade9a98a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.624099] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 903.624099] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bd130-d395-6113-4129-c55e9f271b3f" [ 903.624099] env[61947]: _type = "Task" [ 903.624099] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.636094] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bd130-d395-6113-4129-c55e9f271b3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.669187] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.413s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.672445] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.319s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.673349] env[61947]: DEBUG nova.objects.instance [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lazy-loading 'resources' on Instance uuid e0eebf59-c84e-4462-b280-d783a04525e2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.773057] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224599, 'name': Rename_Task, 'duration_secs': 0.228099} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.773481] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 903.773773] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ed1f281-863d-4540-aee0-6dcfff3b90b0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.781989] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 903.781989] env[61947]: value = "task-1224600" [ 903.781989] env[61947]: _type = "Task" [ 903.781989] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.791455] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224600, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.964150] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224586, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.803055} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.964492] env[61947]: INFO nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899/OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899.vmdk to [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk. [ 903.964756] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Cleaning up location [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 903.964964] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_91ed1f15-2a0a-4e40-9d97-ec8eafd09899 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.965393] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d3885be-9d7c-4424-bf50-fce661714cf8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.974079] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 903.974079] env[61947]: value = "task-1224601" [ 903.974079] env[61947]: _type = "Task" [ 903.974079] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.984318] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.012051] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224594, 'name': PowerOffVM_Task, 'duration_secs': 1.118061} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.014517] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 904.014707] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 904.015014] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-652ccf39-d9a9-417c-8346-6dae8b762e07 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.117304] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 904.117533] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 904.117690] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleting the datastore file [datastore1] 4f313bb5-5df4-4fb0-9736-95f4a519651d {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.117935] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e81b502-8ebd-4cbb-9383-c26a4d9b944a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.130242] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for the task: (returnval){ [ 904.130242] env[61947]: value = "task-1224603" [ 904.130242] env[61947]: _type = "Task" [ 904.130242] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.138242] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bd130-d395-6113-4129-c55e9f271b3f, 'name': SearchDatastore_Task, 'duration_secs': 0.046592} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.138917] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.139202] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.139496] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.139616] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.139839] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.140269] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cad99faa-463b-42e0-8f83-985b0cc723a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.145550] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.149850] env[61947]: DEBUG nova.network.neutron [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updated VIF entry in instance network info cache for port 0b7626fb-f420-4ad3-bea5-09a0daafed10. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.150228] env[61947]: DEBUG nova.network.neutron [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.154295] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.154491] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 904.155352] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-959255cc-051b-4f68-9b6f-51e646b1a873 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.161769] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 904.161769] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a5ffa-6492-17fa-6322-36f9ae00d5ee" [ 904.161769] env[61947]: _type = "Task" [ 904.161769] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.172504] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a5ffa-6492-17fa-6322-36f9ae00d5ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.236807] env[61947]: INFO nova.scheduler.client.report [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocation for migration 91d49010-544b-497e-ad54-e402ab5bdd95 [ 904.297099] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224600, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.303558] env[61947]: DEBUG nova.compute.manager [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-changed-4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.303852] env[61947]: DEBUG nova.compute.manager [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing instance network info cache due to event network-changed-4e1cb946-7339-4abc-a85d-28eb52415eb9. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.304085] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.304276] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.304482] env[61947]: DEBUG nova.network.neutron [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing network info cache for port 4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.451421] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fd88a3-190d-4361-9681-f915da9cb70e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.460502] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a1b5b4-0fd7-4cf5-963d-a62245d989ed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.496924] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac63f1b-1bac-4129-8794-6b39d61c9314 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.508839] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e3ab25-6874-471d-8ba5-815752543552 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.512618] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038512} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.512906] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.513095] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.513348] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk to [datastore1] f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b/f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.513968] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e413920-70dc-401a-9879-1b78e1e39a79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.524060] env[61947]: DEBUG nova.compute.provider_tree [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.528291] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 904.528291] env[61947]: value = "task-1224604" [ 904.528291] env[61947]: _type = "Task" [ 904.528291] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.535687] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.643326] env[61947]: DEBUG oslo_vmware.api [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Task: {'id': task-1224603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159361} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.643660] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.643903] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 904.644119] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 904.644302] env[61947]: INFO nova.compute.manager [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Took 2.18 seconds to destroy the instance on the hypervisor. [ 904.644554] env[61947]: DEBUG oslo.service.loopingcall [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.644779] env[61947]: DEBUG nova.compute.manager [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.644890] env[61947]: DEBUG nova.network.neutron [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.653677] env[61947]: DEBUG oslo_concurrency.lockutils [req-7038c0f5-a25d-4b41-844b-82f1a0105461 req-d3fdb34a-bd22-4c7a-a453-441283068900 service nova] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.674894] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528a5ffa-6492-17fa-6322-36f9ae00d5ee, 'name': SearchDatastore_Task, 'duration_secs': 0.009659} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.675783] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f271e876-23e2-4432-b8e5-4286a966a714 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.684271] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 904.684271] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dfc19f-5fa9-e225-9854-de717258ae96" [ 904.684271] env[61947]: _type = "Task" [ 904.684271] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.694917] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dfc19f-5fa9-e225-9854-de717258ae96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.748418] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7e713153-5a0a-4af8-8307-4e8716dbccec tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.420s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.797622] env[61947]: DEBUG oslo_vmware.api [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224600, 'name': PowerOnVM_Task, 'duration_secs': 0.611193} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.797988] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.798302] env[61947]: INFO nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Took 9.23 seconds to spawn the instance on the hypervisor. [ 904.798574] env[61947]: DEBUG nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.799555] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4819e4-0597-4338-9e4e-9dd8c105da04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.029010] env[61947]: DEBUG nova.scheduler.client.report [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.050777] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.201686] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dfc19f-5fa9-e225-9854-de717258ae96, 'name': SearchDatastore_Task, 'duration_secs': 0.017062} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.201924] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.202217] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 21bb0270-bc20-4ec1-9599-d676845b0dc7/21bb0270-bc20-4ec1-9599-d676845b0dc7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 905.202713] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96308a55-28e7-47cd-8762-21e4948e7db9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.214795] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 905.214795] env[61947]: value = "task-1224605" [ 905.214795] env[61947]: _type = "Task" [ 905.214795] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.231686] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.251830] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 905.251984] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264684', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'name': 'volume-51d58e15-9bd1-43c0-a542-8210c2873e98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd84b03e-ef6d-41e0-b48c-b8e8fc425a36', 'attached_at': '', 'detached_at': '', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'serial': '51d58e15-9bd1-43c0-a542-8210c2873e98'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 905.253243] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ad2ad8-6200-4301-86d3-a63bb303d3a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.284273] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b9ca43-69d8-4344-9623-44412df8284b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.322316] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] volume-51d58e15-9bd1-43c0-a542-8210c2873e98/volume-51d58e15-9bd1-43c0-a542-8210c2873e98.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.327574] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38390fab-9947-49b1-8017-e9bff5391751 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.347150] env[61947]: INFO nova.compute.manager [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Took 24.38 seconds to build instance. [ 905.357136] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 905.357136] env[61947]: value = "task-1224606" [ 905.357136] env[61947]: _type = "Task" [ 905.357136] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.358238] env[61947]: DEBUG nova.network.neutron [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updated VIF entry in instance network info cache for port 4e1cb946-7339-4abc-a85d-28eb52415eb9. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.358681] env[61947]: DEBUG nova.network.neutron [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "address": "fa:16:3e:2f:9a:dc", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1cb946-73", "ovs_interfaceid": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.376626] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224606, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.545214] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.870s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.562025] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.054s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.562517] env[61947]: DEBUG nova.objects.instance [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lazy-loading 'pci_requests' on Instance uuid fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.564388] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.584946] env[61947]: INFO nova.scheduler.client.report [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Deleted allocations for instance e0eebf59-c84e-4462-b280-d783a04525e2 [ 905.732984] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.851041] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f22b6d54-5e76-4747-a28b-028cba109f29 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.900s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.868099] env[61947]: DEBUG oslo_concurrency.lockutils [req-bd48605e-f2e4-4790-a4fe-3b97c96e1d53 req-64e8384b-f486-49a3-914d-a2ff5fc2c384 service nova] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.875972] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.988764] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-7ef85f03-b199-4718-b501-0b5978b80fe0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.989277] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-7ef85f03-b199-4718-b501-0b5978b80fe0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.989859] env[61947]: DEBUG nova.objects.instance [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'flavor' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.048395] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.067661] env[61947]: DEBUG nova.objects.instance [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lazy-loading 'numa_topology' on Instance uuid fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.074883] env[61947]: DEBUG nova.compute.manager [req-50ea1ab9-dea0-4e24-b1da-c538d3648f63 req-acce564e-040a-4f9f-88ce-f01fd7bb415d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-vif-deleted-6832c702-f2d4-446f-a6e3-9225cf51f491 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.075120] env[61947]: INFO nova.compute.manager [req-50ea1ab9-dea0-4e24-b1da-c538d3648f63 req-acce564e-040a-4f9f-88ce-f01fd7bb415d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Neutron deleted interface 6832c702-f2d4-446f-a6e3-9225cf51f491; detaching it from the instance and deleting it from the info cache [ 906.075404] env[61947]: DEBUG nova.network.neutron [req-50ea1ab9-dea0-4e24-b1da-c538d3648f63 req-acce564e-040a-4f9f-88ce-f01fd7bb415d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updating instance_info_cache with network_info: [{"id": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "address": "fa:16:3e:6e:4d:bc", "network": {"id": "b3f69232-c6a8-46da-a6c6-d9d3ec2ae81a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-522940052", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.179", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "025e8ea7833e47e49562ef9ea3bb50a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b59bc8-bc", "ovs_interfaceid": "73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.096291] env[61947]: DEBUG oslo_concurrency.lockutils [None req-21b74eb3-19fa-449a-8192-f021b2043d4a tempest-ServerShowV257Test-202868266 tempest-ServerShowV257Test-202868266-project-member] Lock "e0eebf59-c84e-4462-b280-d783a04525e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.596s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.231054] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.375535] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224606, 'name': ReconfigVM_Task, 'duration_secs': 0.701935} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.375873] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfigured VM instance instance-00000040 to attach disk [datastore1] volume-51d58e15-9bd1-43c0-a542-8210c2873e98/volume-51d58e15-9bd1-43c0-a542-8210c2873e98.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.383755] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e10582c5-53d5-45e5-b92c-45db07dd3910 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.419976] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 906.419976] env[61947]: value = "task-1224607" [ 906.419976] env[61947]: _type = "Task" [ 906.419976] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.433983] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224607, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.508451] env[61947]: DEBUG nova.network.neutron [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.551828] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.583956] env[61947]: INFO nova.compute.claims [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.583956] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4941ba8a-1a86-4072-940b-36434ab47d2e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.600298] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b14f18-1302-4a32-92df-c3760f569b1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.638824] env[61947]: DEBUG nova.compute.manager [req-50ea1ab9-dea0-4e24-b1da-c538d3648f63 req-acce564e-040a-4f9f-88ce-f01fd7bb415d service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Detach interface failed, port_id=6832c702-f2d4-446f-a6e3-9225cf51f491, reason: Instance 4f313bb5-5df4-4fb0-9736-95f4a519651d could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 906.731278] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224605, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.760161] env[61947]: DEBUG nova.objects.instance [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'pci_requests' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.928776] env[61947]: DEBUG oslo_vmware.api [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224607, 'name': ReconfigVM_Task, 'duration_secs': 0.211446} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.929180] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264684', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'name': 'volume-51d58e15-9bd1-43c0-a542-8210c2873e98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd84b03e-ef6d-41e0-b48c-b8e8fc425a36', 'attached_at': '', 'detached_at': '', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'serial': '51d58e15-9bd1-43c0-a542-8210c2873e98'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 907.013290] env[61947]: INFO nova.compute.manager [-] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Took 2.37 seconds to deallocate network for instance. [ 907.050814] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224604, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.40077} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.051126] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/13296250-6feb-41b0-b80f-26c6dd557dde/13296250-6feb-41b0-b80f-26c6dd557dde.vmdk to [datastore1] f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b/f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.052055] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb17355-2a77-4f62-9f41-f89c0c56228e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.081494] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b/f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.082334] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08f799ab-a6c2-40c5-b968-59418a92e227 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.117588] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 907.117588] env[61947]: value = "task-1224608" [ 907.117588] env[61947]: _type = "Task" [ 907.117588] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.136705] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224608, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.024659] env[61947]: DEBUG nova.objects.base [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Object Instance<04e5cc55-71d9-4d3d-95c1-fb1401ab74f8> lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 908.024659] env[61947]: DEBUG nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.028222] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.031138] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224605, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.006717} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.038118] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 21bb0270-bc20-4ec1-9599-d676845b0dc7/21bb0270-bc20-4ec1-9599-d676845b0dc7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 908.038118] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.041958] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e57eeeee-7d8d-4d7a-85e3-28a6c85792c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.053326] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224608, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.054992] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 908.054992] env[61947]: value = "task-1224609" [ 908.054992] env[61947]: _type = "Task" [ 908.054992] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.067118] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.162853] env[61947]: DEBUG nova.policy [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 908.300867] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a710876-954d-40eb-b428-fa18b811f351 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.309782] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3600f6-ea5e-48cb-bfaa-5a360b4d9e2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.349783] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae83f484-178e-45e9-812f-fcdfd11b3aa5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.361213] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7241d4a-c2dc-47ca-9a40-1c2ecd8ee691 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.377390] env[61947]: DEBUG nova.compute.provider_tree [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.542426] env[61947]: INFO nova.compute.manager [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Rescuing [ 908.542992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.542992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.543111] env[61947]: DEBUG nova.network.neutron [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 908.544708] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224608, 'name': ReconfigVM_Task, 'duration_secs': 1.331877} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.546677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.546677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.546677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.546677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.546677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.551049] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Reconfigured VM instance instance-00000049 to attach disk [datastore1] f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b/f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.551049] env[61947]: INFO nova.compute.manager [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Terminating instance [ 908.551049] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-281d01fb-cdc7-43c2-a24d-d896ad9675e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.553872] env[61947]: DEBUG nova.compute.manager [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.553872] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 908.555703] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbde1af-4800-459d-9dea-008e4eb41e71 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.562232] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 908.562232] env[61947]: value = "task-1224610" [ 908.562232] env[61947]: _type = "Task" [ 908.562232] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.574102] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 908.574102] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068334} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.574102] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5e4353b-006b-47cd-9199-41eaa6ec48f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.574240] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.575179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4532c5-e38f-48b4-965c-8b4671132837 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.582634] env[61947]: DEBUG nova.objects.instance [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.583805] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224610, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.587182] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 908.587182] env[61947]: value = "task-1224611" [ 908.587182] env[61947]: _type = "Task" [ 908.587182] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.609869] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 21bb0270-bc20-4ec1-9599-d676845b0dc7/21bb0270-bc20-4ec1-9599-d676845b0dc7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.611072] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e45f2d14-4988-4d77-9860-97da71144f2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.629853] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.636607] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 908.636607] env[61947]: value = "task-1224612" [ 908.636607] env[61947]: _type = "Task" [ 908.636607] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.646166] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.880531] env[61947]: DEBUG nova.scheduler.client.report [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.985889] env[61947]: DEBUG nova.compute.manager [req-fa085403-63ed-47c1-8e69-f2af43cdcf28 req-1987ef77-99e3-4ed4-8880-d6d80804efaa service nova] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Received event network-vif-deleted-73b59bc8-bc5d-4e30-bf7c-6ca0e1bbc4d1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.073443] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224610, 'name': Rename_Task, 'duration_secs': 0.26811} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.075796] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.076057] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e854c6f-d415-4ad2-abf3-b2461dde779e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.084206] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 909.084206] env[61947]: value = "task-1224613" [ 909.084206] env[61947]: _type = "Task" [ 909.084206] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.090045] env[61947]: DEBUG oslo_concurrency.lockutils [None req-25bae19e-19f0-4a78-9366-f601d5dacd24 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.464s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.103401] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224611, 'name': PowerOffVM_Task, 'duration_secs': 0.281733} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.105341] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.105660] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.106041] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.107014] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46058cf6-0ea4-444b-9eec-4e61a38f4458 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.150366] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.186237] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 909.186540] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 909.186798] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleting the datastore file [datastore1] eede4f34-5ac4-475c-a74b-a98327f648fd {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.187153] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82acdfdf-031d-41b3-b0b0-2ff2ac523701 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.195869] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 909.195869] env[61947]: value = "task-1224615" [ 909.195869] env[61947]: _type = "Task" [ 909.195869] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.208403] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.264607] env[61947]: DEBUG nova.network.neutron [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.385878] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.826s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.388134] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 10.502s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.440319] env[61947]: INFO nova.network.neutron [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating port f6236fb9-a6fe-4b8d-948a-0518bada02e3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 909.486776] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.487520] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.596126] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224613, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.650361] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224612, 'name': ReconfigVM_Task, 'duration_secs': 0.613262} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.650680] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 21bb0270-bc20-4ec1-9599-d676845b0dc7/21bb0270-bc20-4ec1-9599-d676845b0dc7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.651394] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cb8178f-1f53-4fb0-8099-485542bddd74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.661105] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 909.661105] env[61947]: value = "task-1224616" [ 909.661105] env[61947]: _type = "Task" [ 909.661105] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.672963] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224616, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.707337] env[61947]: DEBUG oslo_vmware.api [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148305} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.707684] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.708166] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 909.708166] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 909.708357] env[61947]: INFO nova.compute.manager [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Took 1.15 seconds to destroy the instance on the hypervisor. [ 909.708683] env[61947]: DEBUG oslo.service.loopingcall [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.708964] env[61947]: DEBUG nova.compute.manager [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.709107] env[61947]: DEBUG nova.network.neutron [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.770384] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.891663] env[61947]: DEBUG nova.objects.instance [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lazy-loading 'migration_context' on Instance uuid 72818e40-624b-4c04-888b-bb622f7f96d7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.990563] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.065367] env[61947]: DEBUG nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Successfully updated port: 7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.098044] env[61947]: DEBUG oslo_vmware.api [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224613, 'name': PowerOnVM_Task, 'duration_secs': 0.73243} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.098402] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.098785] env[61947]: INFO nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Took 16.99 seconds to spawn the instance on the hypervisor. [ 910.099031] env[61947]: DEBUG nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.099892] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b816b542-d729-4da4-8608-a6fe8d4cd6e1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.180150] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224616, 'name': Rename_Task, 'duration_secs': 0.181794} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.180150] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.180150] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f2c621b-0601-40e6-b290-f69eee0f99bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.190024] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 910.190024] env[61947]: value = "task-1224617" [ 910.190024] env[61947]: _type = "Task" [ 910.190024] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.200309] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.314024] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.314024] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bad075e5-50cf-4e6d-9a79-c50096127686 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.320381] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 910.320381] env[61947]: value = "task-1224618" [ 910.320381] env[61947]: _type = "Task" [ 910.320381] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.331912] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224618, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.365702] env[61947]: DEBUG oslo_concurrency.lockutils [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.366155] env[61947]: DEBUG oslo_concurrency.lockutils [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.366387] env[61947]: DEBUG nova.compute.manager [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.367381] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2948ce0f-9638-4a01-9482-bc8688ecd9b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.375799] env[61947]: DEBUG nova.compute.manager [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 910.376552] env[61947]: DEBUG nova.objects.instance [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.515434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.571985] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.572229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.572425] env[61947]: DEBUG nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 910.619406] env[61947]: DEBUG nova.compute.manager [req-2e9da7bf-5e84-4d25-a20d-de5671838842 req-c417332e-db99-4499-8fed-acdfd8e36a27 service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Received event network-vif-deleted-5e37697a-486f-4961-aed8-8ccdbf59992b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.619705] env[61947]: INFO nova.compute.manager [req-2e9da7bf-5e84-4d25-a20d-de5671838842 req-c417332e-db99-4499-8fed-acdfd8e36a27 service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Neutron deleted interface 5e37697a-486f-4961-aed8-8ccdbf59992b; detaching it from the instance and deleting it from the info cache [ 910.619911] env[61947]: DEBUG nova.network.neutron [req-2e9da7bf-5e84-4d25-a20d-de5671838842 req-c417332e-db99-4499-8fed-acdfd8e36a27 service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.625286] env[61947]: INFO nova.compute.manager [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Took 32.99 seconds to build instance. [ 910.649491] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5776d430-713f-426c-b536-0efa551a3645 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.658749] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0373f8-9ed0-44a7-9185-98f3136a3ad9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.665602] env[61947]: DEBUG nova.network.neutron [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.692999] env[61947]: INFO nova.compute.manager [-] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Took 0.98 seconds to deallocate network for instance. [ 910.696771] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4610ad3-92ef-464b-8576-5057899840ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.710059] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224617, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.711735] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc30b75-6f16-4dfc-a38c-139515a20269 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.728021] env[61947]: DEBUG nova.compute.provider_tree [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.831718] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224618, 'name': PowerOffVM_Task, 'duration_secs': 0.222049} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.832010] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.833070] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6843f9-50aa-4253-85a4-66e29c8416f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.852836] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdac92d-b279-467e-9f7f-9c9a3fdc1325 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.886028] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.886028] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2e8fca2-e2e5-4567-a7a9-3018b56b2cd3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.889197] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.889509] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-944c6969-d064-4a27-a759-66d29e64c151 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.893017] env[61947]: DEBUG oslo_vmware.api [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 910.893017] env[61947]: value = "task-1224619" [ 910.893017] env[61947]: _type = "Task" [ 910.893017] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.901404] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 910.901404] env[61947]: value = "task-1224620" [ 910.901404] env[61947]: _type = "Task" [ 910.901404] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.908869] env[61947]: DEBUG oslo_vmware.api [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.916059] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 910.916336] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.916603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.916866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.917093] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.917404] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8c7ed28-438d-4fa2-8dd2-d2a81dff8b1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.930543] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.930821] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.931634] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcfca980-04ac-4195-a463-d54fa0cd9cd7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.939363] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 910.939363] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc9988-4300-36d0-20fa-be992a8e0202" [ 910.939363] env[61947]: _type = "Task" [ 910.939363] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.948937] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc9988-4300-36d0-20fa-be992a8e0202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.056944] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.057335] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.057571] env[61947]: DEBUG nova.network.neutron [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.117700] env[61947]: WARNING nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] 720a7469-85cf-4e25-82f3-b574730aae1f already exists in list: networks containing: ['720a7469-85cf-4e25-82f3-b574730aae1f']. ignoring it [ 911.117993] env[61947]: WARNING nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] 720a7469-85cf-4e25-82f3-b574730aae1f already exists in list: networks containing: ['720a7469-85cf-4e25-82f3-b574730aae1f']. ignoring it [ 911.127340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-31082ee5-2b4b-4157-83d7-9d0eaee31241 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.505s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.127729] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-679c7022-a13d-46e1-9302-f01e3ac66dc1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.138609] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d00362-6b70-4542-ab95-621f8d6cbcfd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.177436] env[61947]: DEBUG nova.compute.manager [req-2e9da7bf-5e84-4d25-a20d-de5671838842 req-c417332e-db99-4499-8fed-acdfd8e36a27 service nova] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Detach interface failed, port_id=5e37697a-486f-4961-aed8-8ccdbf59992b, reason: Instance eede4f34-5ac4-475c-a74b-a98327f648fd could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 911.202553] env[61947]: DEBUG oslo_vmware.api [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224617, 'name': PowerOnVM_Task, 'duration_secs': 0.796829} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.202869] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.203155] env[61947]: INFO nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Took 10.86 seconds to spawn the instance on the hypervisor. [ 911.203354] env[61947]: DEBUG nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.204156] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f596aadf-a0c7-42c3-925d-cfb6d10fc4b4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.207580] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.234094] env[61947]: DEBUG nova.scheduler.client.report [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.406581] env[61947]: DEBUG oslo_vmware.api [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224619, 'name': PowerOffVM_Task, 'duration_secs': 0.327964} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.407087] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 911.407087] env[61947]: DEBUG nova.compute.manager [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.408295] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bea91ad-6d05-4a91-98a7-e16090d3b158 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.450863] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc9988-4300-36d0-20fa-be992a8e0202, 'name': SearchDatastore_Task, 'duration_secs': 0.010994} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.451679] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-542b3914-45b5-4283-899f-c2eac0cd9e09 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.457546] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 911.457546] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521a5ba4-232c-27c1-a986-0db969b9ad6e" [ 911.457546] env[61947]: _type = "Task" [ 911.457546] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.469686] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521a5ba4-232c-27c1-a986-0db969b9ad6e, 'name': SearchDatastore_Task, 'duration_secs': 0.009512} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.469963] env[61947]: DEBUG oslo_concurrency.lockutils [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.470246] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/24a0d126-58d4-4780-b581-953c5b860675-rescue.vmdk. {{(pid=61947) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 911.470522] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60f29509-3046-497f-8814-2be9bc0b4d05 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.474650] env[61947]: DEBUG nova.compute.manager [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-plugged-7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.474878] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.475109] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.475271] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.475444] env[61947]: DEBUG nova.compute.manager [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] No waiting events found dispatching network-vif-plugged-7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.475611] env[61947]: WARNING nova.compute.manager [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received unexpected event network-vif-plugged-7ef85f03-b199-4718-b501-0b5978b80fe0 for instance with vm_state active and task_state None. [ 911.475778] env[61947]: DEBUG nova.compute.manager [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-changed-7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.475939] env[61947]: DEBUG nova.compute.manager [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing instance network info cache due to event network-changed-7ef85f03-b199-4718-b501-0b5978b80fe0. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.476123] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.481940] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 911.481940] env[61947]: value = "task-1224621" [ 911.481940] env[61947]: _type = "Task" [ 911.481940] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.493881] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.603448] env[61947]: DEBUG nova.network.neutron [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "address": "fa:16:3e:2f:9a:dc", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1cb946-73", "ovs_interfaceid": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ef85f03-b199-4718-b501-0b5978b80fe0", "address": "fa:16:3e:c7:85:04", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ef85f03-b1", "ovs_interfaceid": "7ef85f03-b199-4718-b501-0b5978b80fe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.729339] env[61947]: INFO nova.compute.manager [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Took 28.17 seconds to build instance. [ 911.825507] env[61947]: DEBUG nova.network.neutron [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.911859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.912222] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.921496] env[61947]: DEBUG oslo_concurrency.lockutils [None req-058a9648-ff49-448f-9150-b117f32fa798 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.994858] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46735} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.995140] env[61947]: INFO nova.virt.vmwareapi.ds_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/24a0d126-58d4-4780-b581-953c5b860675-rescue.vmdk. [ 911.996154] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937334d5-43f6-42b0-b310-d78fc656ef19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.034070] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/24a0d126-58d4-4780-b581-953c5b860675-rescue.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.035868] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7e8e34d-34db-4f80-8aa3-d1167618b755 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.062388] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 912.062388] env[61947]: value = "task-1224622" [ 912.062388] env[61947]: _type = "Task" [ 912.062388] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.074543] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224622, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.106729] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.107591] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.107759] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.108111] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.108305] env[61947]: DEBUG nova.network.neutron [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Refreshing network info cache for port 7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.110466] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517abc41-9a48-433c-96b3-cb343ab8e258 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.132855] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.133306] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.133453] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.133740] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.133955] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.134148] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.134389] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.134588] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.134813] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.135053] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.135316] env[61947]: DEBUG nova.virt.hardware [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.142564] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfiguring VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 912.143545] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff5c0806-2b99-4424-adb1-aba3f6647385 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.165429] env[61947]: DEBUG oslo_vmware.api [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 912.165429] env[61947]: value = "task-1224623" [ 912.165429] env[61947]: _type = "Task" [ 912.165429] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.175072] env[61947]: DEBUG oslo_vmware.api [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224623, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.231905] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fe5c587f-0224-47de-aad9-f651f8ea27f2 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.679s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.243749] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.856s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.249972] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.965s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.251683] env[61947]: INFO nova.compute.claims [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.329070] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.356934] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b5628d0305b43637c98bc534e0483b18',container_format='bare',created_at=2024-10-09T14:50:09Z,direct_url=,disk_format='vmdk',id=ef8d583a-38fd-4d74-9ed7-f1718e80cace,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-46796476-shelved',owner='f992ac12ce5244499efac2693a8a517d',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-09T14:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.357272] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.357566] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.357815] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.357992] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.358213] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.358551] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.358744] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.359018] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.359230] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.359416] env[61947]: DEBUG nova.virt.hardware [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.360501] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc6fcc0-aaf8-4c41-b0f8-816d03affdb7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.370696] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcb2a42-10b9-4d3b-9b19-7011c08ef2c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.385530] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:73:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4c8c8fd-baca-4e60-97dc-ff0418d63215', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6236fb9-a6fe-4b8d-948a-0518bada02e3', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.393293] env[61947]: DEBUG oslo.service.loopingcall [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.393948] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.394206] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e4ae095-fab4-4699-a867-e34678f8372e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.415405] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.418171] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.418171] env[61947]: value = "task-1224624" [ 912.418171] env[61947]: _type = "Task" [ 912.418171] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.426964] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224624, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.477043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.477354] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.477647] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.477874] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.478090] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.480906] env[61947]: INFO nova.compute.manager [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Terminating instance [ 912.482755] env[61947]: DEBUG nova.compute.manager [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 912.482921] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 912.483786] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fba933-cd8c-48d2-8623-0505f4b44b64 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.492552] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 912.492845] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1611f6f3-5c9a-4cbb-92a3-8cb52256b2e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.500617] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 912.500617] env[61947]: value = "task-1224625" [ 912.500617] env[61947]: _type = "Task" [ 912.500617] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.511213] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.575446] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.676829] env[61947]: DEBUG oslo_vmware.api [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.892403] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 912.893457] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bde92dc-a7ee-421a-94da-7d8ef3b548e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.900558] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 912.900737] env[61947]: ERROR oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk due to incomplete transfer. [ 912.900978] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8bfb5b3b-ba03-44d1-b65e-bae03dd14dc7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.909868] env[61947]: DEBUG oslo_vmware.rw_handles [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d8bb2-b8f1-04df-565f-ec637ad6418f/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 912.910092] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Uploaded image e4fc5b9a-582d-4cb5-b647-324f95b10a34 to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 912.912418] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 912.912715] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-389d453d-e48b-499f-afa5-5ef6c5600090 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.922126] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 912.922126] env[61947]: value = "task-1224626" [ 912.922126] env[61947]: _type = "Task" [ 912.922126] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.944438] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224624, 'name': CreateVM_Task, 'duration_secs': 0.472212} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.944798] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224626, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.945070] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.945774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.945894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.946420] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.946714] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1eb677a-565c-4426-9742-ad43e376dce6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.949683] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.953982] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 912.953982] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5207f7af-a388-d72d-7228-950fee912dd6" [ 912.953982] env[61947]: _type = "Task" [ 912.953982] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.962866] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5207f7af-a388-d72d-7228-950fee912dd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.011635] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224625, 'name': PowerOffVM_Task, 'duration_secs': 0.361221} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.011910] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.012103] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.012361] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfb1f6e2-eb9d-41bf-b342-307c51faa1a2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.074759] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224622, 'name': ReconfigVM_Task, 'duration_secs': 0.721855} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.075076] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb/24a0d126-58d4-4780-b581-953c5b860675-rescue.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.075976] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bda7b24-5ac5-41c7-a643-d5e869243a3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.106759] env[61947]: DEBUG nova.network.neutron [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updated VIF entry in instance network info cache for port 7ef85f03-b199-4718-b501-0b5978b80fe0. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.107296] env[61947]: DEBUG nova.network.neutron [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "address": "fa:16:3e:2f:9a:dc", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1cb946-73", "ovs_interfaceid": "4e1cb946-7339-4abc-a85d-28eb52415eb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ef85f03-b199-4718-b501-0b5978b80fe0", "address": "fa:16:3e:c7:85:04", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ef85f03-b1", "ovs_interfaceid": "7ef85f03-b199-4718-b501-0b5978b80fe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.108512] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b7a6c3c-9e93-418e-8e4a-f9804b15b444 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.120576] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 913.120822] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 913.121038] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleting the datastore file [datastore1] f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.123059] env[61947]: DEBUG nova.compute.manager [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.123059] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.123228] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.123552] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.123552] env[61947]: DEBUG nova.compute.manager [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] No waiting events found dispatching network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.123666] env[61947]: WARNING nova.compute.manager [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received unexpected event network-vif-plugged-f6236fb9-a6fe-4b8d-948a-0518bada02e3 for instance with vm_state shelved_offloaded and task_state spawning. [ 913.124431] env[61947]: DEBUG nova.compute.manager [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.124431] env[61947]: DEBUG nova.compute.manager [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing instance network info cache due to event network-changed-f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.124431] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Acquiring lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.124431] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Acquired lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.124431] env[61947]: DEBUG nova.network.neutron [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Refreshing network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.127036] env[61947]: DEBUG oslo_concurrency.lockutils [req-90388c7a-4a7a-4c02-b5fa-6fb7d8b0b60f req-ee039c8c-ca6d-4a6c-96ca-d073dec18fdf service nova] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.127390] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00de09f1-d63e-4bf4-9e99-7f447a35f009 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.136803] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 913.136803] env[61947]: value = "task-1224628" [ 913.136803] env[61947]: _type = "Task" [ 913.136803] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.138830] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 913.138830] env[61947]: value = "task-1224629" [ 913.138830] env[61947]: _type = "Task" [ 913.138830] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.153258] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224629, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.157425] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224628, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.176705] env[61947]: DEBUG oslo_vmware.api [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224623, 'name': ReconfigVM_Task, 'duration_secs': 0.869566} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.177250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.177475] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfigured VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 913.441896] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224626, 'name': Destroy_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.466292] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.466434] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Processing image ef8d583a-38fd-4d74-9ed7-f1718e80cace {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.466669] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.466823] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.467037] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.467288] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4d5b0be-5bfc-470b-9be6-733b8a2c3977 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.479109] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.479276] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.480120] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-831c693f-4319-41ca-bff2-7319f869d8cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.488599] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 913.488599] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b3896d-193a-3221-1629-9a41afdfab96" [ 913.488599] env[61947]: _type = "Task" [ 913.488599] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.497233] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b3896d-193a-3221-1629-9a41afdfab96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.498855] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f36abc-e2b1-49fc-a1d6-035671758f04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.506127] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db48489-95db-4991-b4f6-4010429dbc97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.536545] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1bf5ff-f862-4607-966e-c005adc29ab4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.544798] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb1ceb5-f1a7-4ce2-bc23-f3dcb73a8c11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.558469] env[61947]: DEBUG nova.compute.provider_tree [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.653285] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224628, 'name': ReconfigVM_Task, 'duration_secs': 0.361162} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.656437] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.656779] env[61947]: DEBUG oslo_vmware.api [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224629, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136644} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.659175] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-278453c0-ea1f-48bf-9557-8b7aa73610db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.660661] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.661289] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 913.661289] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 913.661289] env[61947]: INFO nova.compute.manager [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Took 1.18 seconds to destroy the instance on the hypervisor. [ 913.661416] env[61947]: DEBUG oslo.service.loopingcall [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.661622] env[61947]: DEBUG nova.compute.manager [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 913.661723] env[61947]: DEBUG nova.network.neutron [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.669088] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 913.669088] env[61947]: value = "task-1224630" [ 913.669088] env[61947]: _type = "Task" [ 913.669088] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.677708] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.682559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3ff91ce9-9208-4525-89eb-22a6c60d7932 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-7ef85f03-b199-4718-b501-0b5978b80fe0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.693s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.717447] env[61947]: DEBUG nova.compute.manager [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Received event network-changed-0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.717703] env[61947]: DEBUG nova.compute.manager [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Refreshing instance network info cache due to event network-changed-0b7626fb-f420-4ad3-bea5-09a0daafed10. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.717915] env[61947]: DEBUG oslo_concurrency.lockutils [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.718268] env[61947]: DEBUG oslo_concurrency.lockutils [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.718486] env[61947]: DEBUG nova.network.neutron [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Refreshing network info cache for port 0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.792249] env[61947]: INFO nova.compute.manager [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Swapping old allocation on dict_keys(['7263fe00-9b30-4433-9e9f-ec265ed2b8f2']) held by migration 89e41994-bff5-4166-a6a9-10258b48d1d3 for instance [ 913.815512] env[61947]: DEBUG nova.scheduler.client.report [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Overwriting current allocation {'allocations': {'7263fe00-9b30-4433-9e9f-ec265ed2b8f2': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 98}}, 'project_id': 'ef226aa2daf94b9eaeb3b6b227292ff4', 'user_id': 'd85ffd46682f45ef9302c6fd233ab4ce', 'consumer_generation': 1} on consumer 72818e40-624b-4c04-888b-bb622f7f96d7 {{(pid=61947) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 913.899340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.899550] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquired lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.899741] env[61947]: DEBUG nova.network.neutron [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.940689] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224626, 'name': Destroy_Task, 'duration_secs': 0.6867} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.940982] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Destroyed the VM [ 913.941285] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 913.942136] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9841f016-4f76-44f3-96f6-141936acc63e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.950454] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 913.950454] env[61947]: value = "task-1224631" [ 913.950454] env[61947]: _type = "Task" [ 913.950454] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.961379] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224631, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.983241] env[61947]: DEBUG nova.network.neutron [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updated VIF entry in instance network info cache for port f6236fb9-a6fe-4b8d-948a-0518bada02e3. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.983678] env[61947]: DEBUG nova.network.neutron [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [{"id": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "address": "fa:16:3e:07:73:45", "network": {"id": "e1b9c58f-84d7-4ac0-87c8-f10b0ff073bb", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-604873078-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f992ac12ce5244499efac2693a8a517d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6236fb9-a6", "ovs_interfaceid": "f6236fb9-a6fe-4b8d-948a-0518bada02e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.001753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 914.002036] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Fetch image to [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37/OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 914.002232] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Downloading stream optimized image ef8d583a-38fd-4d74-9ed7-f1718e80cace to [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37/OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37.vmdk on the data store datastore1 as vApp {{(pid=61947) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 914.002486] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Downloading image file data ef8d583a-38fd-4d74-9ed7-f1718e80cace to the ESX as VM named 'OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37' {{(pid=61947) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 914.062929] env[61947]: DEBUG nova.scheduler.client.report [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.082506] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 914.082506] env[61947]: value = "resgroup-9" [ 914.082506] env[61947]: _type = "ResourcePool" [ 914.082506] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 914.082846] env[61947]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-19a947d1-326b-4761-b39e-aa59d59f436f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.105416] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lease: (returnval){ [ 914.105416] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 914.105416] env[61947]: _type = "HttpNfcLease" [ 914.105416] env[61947]: } obtained for vApp import into resource pool (val){ [ 914.105416] env[61947]: value = "resgroup-9" [ 914.105416] env[61947]: _type = "ResourcePool" [ 914.105416] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 914.105750] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the lease: (returnval){ [ 914.105750] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 914.105750] env[61947]: _type = "HttpNfcLease" [ 914.105750] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 914.113352] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.113352] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 914.113352] env[61947]: _type = "HttpNfcLease" [ 914.113352] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 914.180484] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224630, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.248753] env[61947]: DEBUG nova.compute.manager [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Stashing vm_state: stopped {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 914.461548] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224631, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.486490] env[61947]: DEBUG oslo_concurrency.lockutils [req-2456b2e2-dcc4-48f4-9416-e726c9e89360 req-5e8c97f9-590f-48ad-80dd-79753db23941 service nova] Releasing lock "refresh_cache-fa0eed6e-831f-4e7a-9890-fc5a02609aa4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.495431] env[61947]: DEBUG nova.network.neutron [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updated VIF entry in instance network info cache for port 0b7626fb-f420-4ad3-bea5-09a0daafed10. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.495768] env[61947]: DEBUG nova.network.neutron [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.556692] env[61947]: DEBUG nova.network.neutron [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.568593] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.568811] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.579840] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.551s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.579840] env[61947]: DEBUG nova.objects.instance [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lazy-loading 'resources' on Instance uuid 4f313bb5-5df4-4fb0-9736-95f4a519651d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.614483] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.614483] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 914.614483] env[61947]: _type = "HttpNfcLease" [ 914.614483] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 914.682727] env[61947]: DEBUG oslo_vmware.api [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224630, 'name': PowerOnVM_Task, 'duration_secs': 0.544719} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.682727] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.684129] env[61947]: DEBUG nova.compute.manager [None req-993996f6-9933-4dcc-b1ac-649203573813 tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.685152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf5b7fd-4983-4938-a9f0-c7f87a7647d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.968524] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224631, 'name': RemoveSnapshot_Task, 'duration_secs': 0.607041} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.968524] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 914.968524] env[61947]: DEBUG nova.compute.manager [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.969353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f372c2d1-8c52-4473-a724-a409f4c902e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.992680] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.001316] env[61947]: DEBUG oslo_concurrency.lockutils [req-35421e0d-80b5-4f87-a7b0-1aba41c6ed81 req-d24b7d21-f357-46cc-9a54-327357015ec8 service nova] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.050904] env[61947]: DEBUG nova.network.neutron [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [{"id": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "address": "fa:16:3e:a8:b8:60", "network": {"id": "5d0bfdf9-e023-47c4-b1d0-58757c2ca53b", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.214", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "7706c941d6d844f588f72bcf7d3ea118", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069a41b-ac", "ovs_interfaceid": "c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.057762] env[61947]: INFO nova.compute.manager [-] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Took 1.40 seconds to deallocate network for instance. [ 915.084501] env[61947]: DEBUG nova.compute.utils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.086082] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.089185] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.117205] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.117205] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 915.117205] env[61947]: _type = "HttpNfcLease" [ 915.117205] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 915.117205] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 915.117205] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5285de95-eda0-2410-0232-bc6e83f13b47" [ 915.117205] env[61947]: _type = "HttpNfcLease" [ 915.117205] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 915.119976] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e7b23f-9613-4425-88b2-b0cd3082fb8f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.131947] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 915.134088] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 915.199569] env[61947]: DEBUG nova.policy [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ae1ce526675412aa854af0c99fb70ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5b54acf7e242414ab95acfe5433ab171', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 915.202294] env[61947]: DEBUG nova.compute.manager [req-108681c0-b0ad-4825-9616-994a452b988b req-6f22763c-9947-45a6-82df-6c92e84765c6 service nova] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Received event network-vif-deleted-9de62525-ad88-46d7-9cab-b679b8515e82 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.208160] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-76d8de98-a58f-4c30-9920-350e73c83cca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.407408] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1da039-535e-4090-a228-abf6ee3c1c58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.417319] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4174237-ca72-47c4-a1e1-4ada92bb7ffa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.448099] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39ff1ec-68ce-4667-97fa-e8bf6c74672c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.456400] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b34b79-3e93-4c47-90ca-16bc099eb86a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.470618] env[61947]: DEBUG nova.compute.provider_tree [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.484674] env[61947]: INFO nova.compute.manager [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Shelve offloading [ 915.486280] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.489250] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d57aaf64-ba97-427a-89f7-84f06da7848b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.495890] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 915.495890] env[61947]: value = "task-1224633" [ 915.495890] env[61947]: _type = "Task" [ 915.495890] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.505062] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.552987] env[61947]: DEBUG oslo_concurrency.lockutils [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Releasing lock "refresh_cache-72818e40-624b-4c04-888b-bb622f7f96d7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.553506] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.553776] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d42c0c25-7929-4361-967b-74a7bb79d4ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.561596] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 915.561596] env[61947]: value = "task-1224634" [ 915.561596] env[61947]: _type = "Task" [ 915.561596] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.565704] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.572851] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.591867] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.680731] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Successfully created port: c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.974319] env[61947]: DEBUG nova.scheduler.client.report [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.013342] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 916.013685] env[61947]: DEBUG nova.compute.manager [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.014650] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a53e7de-7eac-45cc-a259-47e3c5bc30c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.025064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.025275] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.026038] env[61947]: DEBUG nova.network.neutron [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.075075] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224634, 'name': PowerOffVM_Task, 'duration_secs': 0.190688} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.077073] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.077866] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:48:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='9a82a007-fbf5-4f42-9bd3-1a8156f71854',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-569079550',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.078118] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.078348] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.078650] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.078829] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.079146] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.079355] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.079564] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.079854] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.080098] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.080305] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.097069] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f33157d-c0a5-4c2a-a27c-c8c484762a8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.110277] env[61947]: INFO nova.virt.block_device [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Booting with volume 4639dbb5-6722-408e-92c6-d142b4cdeb3b at /dev/sda [ 916.128961] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 916.128961] env[61947]: value = "task-1224635" [ 916.128961] env[61947]: _type = "Task" [ 916.128961] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.140531] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224635, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.163616] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9851f2b9-a014-431d-a860-a15599e750da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.175509] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc659330-fdd3-4da8-84f1-a14584583b01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.216817] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2811af2a-bd04-4af9-a594-bde85e5fe813 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.219793] env[61947]: DEBUG nova.compute.manager [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.220221] env[61947]: DEBUG nova.compute.manager [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing instance network info cache due to event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 916.220517] env[61947]: DEBUG oslo_concurrency.lockutils [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.220733] env[61947]: DEBUG oslo_concurrency.lockutils [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.220989] env[61947]: DEBUG nova.network.neutron [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.232349] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8586b62a-7a11-4aee-806f-edba6867015f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.276939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-4e1cb946-7339-4abc-a85d-28eb52415eb9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.277190] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-4e1cb946-7339-4abc-a85d-28eb52415eb9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.284290] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47163b5-8f33-41da-8823-48c14319584f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.294382] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d161650-f35d-4dd8-bf99-d83ba41c52a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.312484] env[61947]: DEBUG nova.virt.block_device [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updating existing volume attachment record: db781476-917f-4463-a528-6e667bdfd975 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 916.482437] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.484799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.970s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.486483] env[61947]: INFO nova.compute.claims [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.506284] env[61947]: INFO nova.scheduler.client.report [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Deleted allocations for instance 4f313bb5-5df4-4fb0-9736-95f4a519651d [ 916.649490] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224635, 'name': ReconfigVM_Task, 'duration_secs': 0.183256} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.653492] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd6d871-95a0-4077-a03f-924f9e1b4b58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.681855] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:48:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='9a82a007-fbf5-4f42-9bd3-1a8156f71854',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-569079550',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.682109] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.682270] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.682451] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.682602] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.682742] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.682946] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.683119] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.683456] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.684122] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.684122] env[61947]: DEBUG nova.virt.hardware [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.686862] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dc8a256-9a37-454e-b8d4-d994ebf78573 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.694815] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 916.694815] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521ff535-cd00-e708-a42a-9b93950636a9" [ 916.694815] env[61947]: _type = "Task" [ 916.694815] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.710427] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521ff535-cd00-e708-a42a-9b93950636a9, 'name': SearchDatastore_Task, 'duration_secs': 0.010648} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.719588] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 916.720065] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb6a932b-6dae-4ce5-8178-1d1d1cbf7b3b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.772239] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 916.772239] env[61947]: value = "task-1224636" [ 916.772239] env[61947]: _type = "Task" [ 916.772239] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.782380] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224636, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.788115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.788305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.789195] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1bdfa7-6cf5-4cd5-8e99-1984c61d1bca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.813088] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaced802-ec91-449b-903f-b00c47f5ff81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.844103] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfiguring VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 916.848310] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9de8add-8834-4758-be7a-71d9004f3058 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.872551] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 916.872551] env[61947]: value = "task-1224637" [ 916.872551] env[61947]: _type = "Task" [ 916.872551] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.883326] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.909762] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 916.910148] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 916.911404] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dff99da-75de-412d-be4f-bd668c70a1ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.920055] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 916.920332] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 916.920668] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9ab586d8-1669-4c58-ab81-4d1010cb12b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.935171] env[61947]: DEBUG nova.network.neutron [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updating instance_info_cache with network_info: [{"id": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "address": "fa:16:3e:18:eb:2c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcfe6b8e-1a", "ovs_interfaceid": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.018895] env[61947]: DEBUG oslo_concurrency.lockutils [None req-35a39633-7015-4748-986d-59251dbc1213 tempest-ServersTestMultiNic-1830921944 tempest-ServersTestMultiNic-1830921944-project-member] Lock "4f313bb5-5df4-4fb0-9736-95f4a519651d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.561s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.063192] env[61947]: DEBUG nova.network.neutron [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updated VIF entry in instance network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.063578] env[61947]: DEBUG nova.network.neutron [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.228302] env[61947]: DEBUG nova.compute.manager [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.228613] env[61947]: DEBUG nova.compute.manager [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing instance network info cache due to event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.228775] env[61947]: DEBUG oslo_concurrency.lockutils [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.283414] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224636, 'name': ReconfigVM_Task, 'duration_secs': 0.281441} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.283743] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 917.284539] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703d3345-1a88-4350-916e-9fe19468f412 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.310709] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.311079] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-511729fe-fff3-4a00-9314-44f1c8034f7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.334814] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 917.334814] env[61947]: value = "task-1224638" [ 917.334814] env[61947]: _type = "Task" [ 917.334814] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.340684] env[61947]: DEBUG oslo_vmware.rw_handles [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528213fb-2150-071f-9bc2-1a1641bea148/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 917.340982] env[61947]: INFO nova.virt.vmwareapi.images [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Downloaded image file data ef8d583a-38fd-4d74-9ed7-f1718e80cace [ 917.342435] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a58ac0-db1c-4ba9-8ec0-1c9c04b21ac0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.350588] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.365468] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-923fa15b-9e3f-4f9f-b776-24fec6e434cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.384916] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.423816] env[61947]: INFO nova.virt.vmwareapi.images [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] The imported VM was unregistered [ 917.426444] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 917.426713] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Creating directory with path [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.426999] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a21a08d-bc40-4a77-8c2d-b7e2433f6dec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.438484] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.442309] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Created directory with path [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.442516] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37/OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37.vmdk to [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk. {{(pid=61947) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 917.442853] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-55a5a393-d76b-443d-bd26-a9b8abf88125 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.450998] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 917.450998] env[61947]: value = "task-1224640" [ 917.450998] env[61947]: _type = "Task" [ 917.450998] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.460177] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.462651] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Successfully updated port: c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.568146] env[61947]: DEBUG oslo_concurrency.lockutils [req-2f6b74f7-cae2-4b43-9035-0cbd7b514b4c req-27a0af2c-e41c-4d98-a7c9-88d6f9a0468c service nova] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.568146] env[61947]: DEBUG oslo_concurrency.lockutils [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.568244] env[61947]: DEBUG nova.network.neutron [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.758831] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d86866-cabf-4860-bdc4-84f1fed66487 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.767455] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00691d27-a413-46c2-b522-62f0955bfd59 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.803410] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4422933-efe0-4f45-bd55-416fd3e52fa0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.814779] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d74891-8aeb-4d39-9b31-dc8f236f6b1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.834775] env[61947]: DEBUG nova.compute.provider_tree [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.849339] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224638, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.887019] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.909062] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 917.909961] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2770978-1fb7-4e67-9f40-1801e51bac17 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.925151] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 917.925460] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1249912-4f13-4c02-bfd8-56a5411646c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.966313] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.966727] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquired lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.966727] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.968997] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.012670] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 918.012948] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 918.013202] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleting the datastore file [datastore2] 9cfa7c52-2250-49ce-a1c8-baad6644f824 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.013505] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb473768-fc9b-4ee5-9afb-39eaaa4675bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.019724] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.019921] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.028443] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 918.028443] env[61947]: value = "task-1224642" [ 918.028443] env[61947]: _type = "Task" [ 918.028443] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.043888] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.339971] env[61947]: DEBUG nova.scheduler.client.report [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.358980] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224638, 'name': ReconfigVM_Task, 'duration_secs': 0.66996} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.359412] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7/72818e40-624b-4c04-888b-bb622f7f96d7.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.360448] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503c578a-28fc-4904-892e-dbc213905913 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.390360] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a6b7e8-8261-4d61-b6dd-9e3a2f6af760 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.424478] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f612e7-6881-4c46-bc4d-bba0d549a140 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.427673] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.428709] env[61947]: DEBUG nova.network.neutron [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updated VIF entry in instance network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.429087] env[61947]: DEBUG nova.network.neutron [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.432039] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.432657] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.432898] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.433076] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.433270] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.433780] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.433955] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.434191] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.434502] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.434739] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.434918] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.435111] env[61947]: DEBUG nova.virt.hardware [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.450301] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea25ff0c-6bd7-47f3-a47e-d37cf195bb44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.454827] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc419fd6-72ca-4d46-9e53-b4aa55167f7e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.471919] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.478534] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3072896-c73b-490b-a87c-55ed3b8c0014 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.480692] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.483353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232585ae-2324-47a4-ae3a-848b9d3a79c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.503919] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 918.503919] env[61947]: value = "task-1224643" [ 918.503919] env[61947]: _type = "Task" [ 918.503919] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.508953] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.517254] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224643, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.523519] env[61947]: DEBUG nova.compute.utils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.548497] env[61947]: DEBUG oslo_vmware.api [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260345} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.548858] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.549090] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 918.549279] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 918.573332] env[61947]: INFO nova.scheduler.client.report [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance 9cfa7c52-2250-49ce-a1c8-baad6644f824 [ 918.665391] env[61947]: DEBUG nova.network.neutron [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updating instance_info_cache with network_info: [{"id": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "address": "fa:16:3e:ae:bb:b9", "network": {"id": "347593d1-862d-4d28-b3ec-328cf05d68f9", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1066216228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b54acf7e242414ab95acfe5433ab171", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2f0d861-74", "ovs_interfaceid": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.847845] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.848516] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.851560] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.644s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.851713] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.853838] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.904s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.856241] env[61947]: INFO nova.compute.claims [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.877914] env[61947]: INFO nova.scheduler.client.report [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocations for instance eede4f34-5ac4-475c-a74b-a98327f648fd [ 918.902556] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.932037] env[61947]: DEBUG oslo_concurrency.lockutils [req-0559ca69-c05a-4821-be94-07acdcd86410 req-3a932fe0-54d3-476f-a26c-237add387f76 service nova] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.971946] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.021489] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224643, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.027195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.042131] env[61947]: DEBUG nova.compute.manager [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received event network-vif-unplugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.042415] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.042580] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.042801] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.042969] env[61947]: DEBUG nova.compute.manager [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] No waiting events found dispatching network-vif-unplugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.043368] env[61947]: WARNING nova.compute.manager [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received unexpected event network-vif-unplugged-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef for instance with vm_state shelved_offloaded and task_state None. [ 919.043569] env[61947]: DEBUG nova.compute.manager [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Received event network-changed-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.043765] env[61947]: DEBUG nova.compute.manager [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Refreshing instance network info cache due to event network-changed-dcfe6b8e-1a14-4344-a03c-1f62c9b7baef. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 919.043929] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Acquiring lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.044083] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Acquired lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.044247] env[61947]: DEBUG nova.network.neutron [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Refreshing network info cache for port dcfe6b8e-1a14-4344-a03c-1f62c9b7baef {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.077943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.168983] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Releasing lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.169434] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance network_info: |[{"id": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "address": "fa:16:3e:ae:bb:b9", "network": {"id": "347593d1-862d-4d28-b3ec-328cf05d68f9", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1066216228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b54acf7e242414ab95acfe5433ab171", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2f0d861-74", "ovs_interfaceid": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.169909] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:bb:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a44713-0af1-486e-bc0d-00e03a769fa4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2f0d861-74c8-4b27-abb4-94bfa51b7768', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.177703] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Creating folder: Project (5b54acf7e242414ab95acfe5433ab171). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 919.178448] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d1b90d2-6db8-4e12-beb0-870d7ef4a54c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.198645] env[61947]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 919.198850] env[61947]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61947) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 919.199270] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Folder already exists: Project (5b54acf7e242414ab95acfe5433ab171). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 919.199479] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Creating folder: Instances. Parent ref: group-v264664. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 919.199757] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2a7d903-c3ff-4e8d-ab6b-928f732906a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.215214] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Created folder: Instances in parent group-v264664. [ 919.215691] env[61947]: DEBUG oslo.service.loopingcall [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.216049] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 919.216411] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0070152-a2e9-4d85-a0d7-8910258a2910 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.243285] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.243285] env[61947]: value = "task-1224646" [ 919.243285] env[61947]: _type = "Task" [ 919.243285] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.256802] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224646, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.360159] env[61947]: DEBUG nova.compute.utils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.363964] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.364204] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.394911] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ddd35861-b37f-419b-b972-c8c62c1a7a0b tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "eede4f34-5ac4-475c-a74b-a98327f648fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.849s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.406886] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.413690] env[61947]: DEBUG nova.policy [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0337baafe604194a5bc93e99ca8078a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51aa57efe97e453783044286f33d3f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 919.472850] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.520980] env[61947]: DEBUG oslo_vmware.api [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224643, 'name': PowerOnVM_Task, 'duration_secs': 0.560379} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.521248] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.750228] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Successfully created port: 3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.778403] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224646, 'name': CreateVM_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.869923] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.910785] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.975280] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.089842] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3fdc22-10e1-4928-9eb6-1d2e2d6b8c85 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.098449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad361999-d8af-432d-ba2a-522408ecd5d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.102298] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.102542] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.102838] env[61947]: INFO nova.compute.manager [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Attaching volume 914190f7-1c29-41db-8c24-20092cdc4528 to /dev/sdb [ 920.108544] env[61947]: DEBUG nova.network.neutron [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updated VIF entry in instance network info cache for port dcfe6b8e-1a14-4344-a03c-1f62c9b7baef. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.108955] env[61947]: DEBUG nova.network.neutron [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updating instance_info_cache with network_info: [{"id": "dcfe6b8e-1a14-4344-a03c-1f62c9b7baef", "address": "fa:16:3e:18:eb:2c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": null, "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdcfe6b8e-1a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.147416] env[61947]: DEBUG oslo_concurrency.lockutils [req-a7fd4354-15c9-4ba3-9f78-15ca1d8b19a5 req-083efbfb-0aa2-4fad-b8cb-65c883c312ad service nova] Releasing lock "refresh_cache-9cfa7c52-2250-49ce-a1c8-baad6644f824" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.148386] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dc1673-2d87-4d9d-afd3-cb2acb6ecf04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.158072] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255a80bb-f46f-4581-afcc-7328687777dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.176440] env[61947]: DEBUG nova.compute.provider_tree [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.189183] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836931a1-f27d-4fe1-926b-a95eb27d002f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.199448] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15cfb5b-f512-4f8e-b1ba-62c75523df57 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.219176] env[61947]: DEBUG nova.virt.block_device [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updating existing volume attachment record: e674e752-91e4-4269-9be0-9dfe155a62bc {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 920.255626] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224646, 'name': CreateVM_Task, 'duration_secs': 0.977481} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.255802] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 920.256519] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'device_type': None, 'mount_device': '/dev/sda', 'attachment_id': 'db781476-917f-4463-a528-6e667bdfd975', 'guest_format': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264671', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'name': 'volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '91205e54-6bcb-403c-8308-0f0692fe4fba', 'attached_at': '', 'detached_at': '', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'serial': '4639dbb5-6722-408e-92c6-d142b4cdeb3b'}, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61947) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 920.256826] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Root volume attach. Driver type: vmdk {{(pid=61947) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 920.257944] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102927f4-5839-4e4d-a876-640f754fbcda {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.268475] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8196eb1f-accb-42d7-ae0e-9386a974ef74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.276038] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d951f52a-1317-4958-8f79-2dae8840effd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.285532] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-41e81432-1f4a-4442-880f-db818ab69e1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.302926] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 920.302926] env[61947]: value = "task-1224647" [ 920.302926] env[61947]: _type = "Task" [ 920.302926] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.307182] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.401293] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.432495] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Received event network-vif-plugged-c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.432605] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Acquiring lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.432785] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.432965] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.435242] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] No waiting events found dispatching network-vif-plugged-c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.435462] env[61947]: WARNING nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Received unexpected event network-vif-plugged-c2f0d861-74c8-4b27-abb4-94bfa51b7768 for instance with vm_state building and task_state spawning. [ 920.435680] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Received event network-changed-c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.435845] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Refreshing instance network info cache due to event network-changed-c2f0d861-74c8-4b27-abb4-94bfa51b7768. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.436092] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Acquiring lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.436317] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Acquired lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.436497] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Refreshing network info cache for port c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.466792] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224640, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.608629} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.467154] env[61947]: INFO nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37/OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37.vmdk to [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk. [ 920.467375] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Cleaning up location [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 920.467556] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e6e19972-32d6-4b52-81e9-556470937b37 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.467830] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7b535d1-90cb-41cd-9a7f-2aaf8013970e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.476221] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 920.476221] env[61947]: value = "task-1224650" [ 920.476221] env[61947]: _type = "Task" [ 920.476221] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.487236] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224650, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.534633] env[61947]: INFO nova.compute.manager [None req-09db9615-e6b0-4634-a720-2b374f758072 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance to original state: 'active' [ 920.678717] env[61947]: DEBUG nova.scheduler.client.report [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.810563] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 42%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.880635] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.906578] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.909486] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.909766] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.909939] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.910145] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.910294] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.910447] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.910660] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.910825] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.910999] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.911405] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.911591] env[61947]: DEBUG nova.virt.hardware [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.912648] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e70e6e1-8b05-40fb-8886-028a941b2949 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.922879] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19963240-5e30-4dff-87d9-231eaebc9f7e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.987248] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224650, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.051495} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.987560] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.987740] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.988014] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk to [datastore1] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 920.988322] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae84fcf7-bd15-4c90-91b8-b136fa3c0282 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.996920] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.998731] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 920.998731] env[61947]: value = "task-1224652" [ 920.998731] env[61947]: _type = "Task" [ 920.998731] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.010266] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.184048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.185351] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.188658] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.196s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.310755] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updated VIF entry in instance network info cache for port c2f0d861-74c8-4b27-abb4-94bfa51b7768. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.311543] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updating instance_info_cache with network_info: [{"id": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "address": "fa:16:3e:ae:bb:b9", "network": {"id": "347593d1-862d-4d28-b3ec-328cf05d68f9", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1066216228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b54acf7e242414ab95acfe5433ab171", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2f0d861-74", "ovs_interfaceid": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.321142] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 54%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.408189] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.512673] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.674642] env[61947]: DEBUG nova.compute.manager [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Received event network-vif-plugged-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 921.675172] env[61947]: DEBUG oslo_concurrency.lockutils [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.675623] env[61947]: DEBUG oslo_concurrency.lockutils [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.675992] env[61947]: DEBUG oslo_concurrency.lockutils [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.677936] env[61947]: DEBUG nova.compute.manager [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] No waiting events found dispatching network-vif-plugged-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.678292] env[61947]: WARNING nova.compute.manager [req-bf890d90-0350-40d8-82ed-c1e91efcfe87 req-76bdbcee-49f4-4572-9667-e47f9489b4aa service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Received unexpected event network-vif-plugged-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 for instance with vm_state building and task_state spawning. [ 921.691906] env[61947]: DEBUG nova.compute.utils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.696571] env[61947]: INFO nova.compute.claims [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.705039] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.705039] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 921.790931] env[61947]: DEBUG nova.policy [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '897eb27f633a4a65bd500448ede60c1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7901e6bbfd1e4832a556d1a2b58d9850', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 921.818145] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Releasing lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.818450] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 921.818699] env[61947]: DEBUG nova.compute.manager [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing instance network info cache due to event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 921.818946] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.819112] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.819534] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.821126] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 69%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.900565] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Successfully updated port: 3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.913143] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.975561] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.975998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.976259] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.976996] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.977215] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.979941] env[61947]: INFO nova.compute.manager [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Terminating instance [ 921.982578] env[61947]: DEBUG nova.compute.manager [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.982806] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 921.983971] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d876d0e9-f525-42fa-a9db-bda183b5899a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.999100] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 921.999100] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ef8ba67-f06a-4cca-9176-7031411d7c83 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.008556] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 922.008556] env[61947]: value = "task-1224653" [ 922.008556] env[61947]: _type = "Task" [ 922.008556] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.012311] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.034255] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.207640] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.213450] env[61947]: INFO nova.compute.resource_tracker [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating resource usage from migration 33113961-17ce-41e5-b039-4af1e078f3ef [ 922.319480] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 84%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.323281] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Successfully created port: c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.404956] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.404956] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.404956] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.413552] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.507213] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256b0800-a5ed-48fe-95e7-b7152c2358a4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.525036] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.525036] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aeb6499-fe94-4ff3-a407-4d6d7a1ebf60 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.531545] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224653, 'name': PowerOffVM_Task, 'duration_secs': 0.392879} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.532426] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 922.532602] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.532974] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c763822-e086-4004-a769-683961d00a28 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.569624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810ecdb5-3c2b-4247-8b37-1f786627d52e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.580325] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41632762-d034-4b6d-9938-5374c1712582 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.602620] env[61947]: DEBUG nova.compute.provider_tree [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.607151] env[61947]: DEBUG nova.compute.manager [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.607384] env[61947]: DEBUG nova.compute.manager [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing instance network info cache due to event network-changed-ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.607589] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] Acquiring lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.618479] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.618817] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.619116] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Deleting the datastore file [datastore1] 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.619488] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8172498-629f-4aa2-a8d0-9bd6a5981460 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.624420] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "72818e40-624b-4c04-888b-bb622f7f96d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.624703] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.624967] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.625243] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.625459] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.628066] env[61947]: INFO nova.compute.manager [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Terminating instance [ 922.631315] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for the task: (returnval){ [ 922.631315] env[61947]: value = "task-1224656" [ 922.631315] env[61947]: _type = "Task" [ 922.631315] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.631980] env[61947]: DEBUG nova.compute.manager [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.632276] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.633542] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27094bad-37e9-404f-8ff8-6380a853ee9d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.649546] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.652435] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.652841] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a1fe3a0-bf91-41ea-a309-f21904f6691b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.663471] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 922.663471] env[61947]: value = "task-1224657" [ 922.663471] env[61947]: _type = "Task" [ 922.663471] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.678476] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.823225] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 95%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.910168] env[61947]: DEBUG oslo_vmware.api [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224637, 'name': ReconfigVM_Task, 'duration_secs': 5.839704} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.910453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.910666] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Reconfigured VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 922.963266] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.017656] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.040677] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updated VIF entry in instance network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.041064] env[61947]: DEBUG nova.network.neutron [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.109629] env[61947]: DEBUG nova.scheduler.client.report [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.154472] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.177303] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224657, 'name': PowerOffVM_Task, 'duration_secs': 0.231221} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.177605] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.177787] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 923.178306] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb3fa7f5-8912-4fbf-bf56-36025f38e01f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.211605] env[61947]: DEBUG nova.network.neutron [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.225708] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.261505] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.261777] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.261940] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.262772] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.263180] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.263959] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.264330] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.264557] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.265051] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.265051] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.265260] env[61947]: DEBUG nova.virt.hardware [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.266942] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86de7e44-e3bb-408c-93ee-32e1e21efa40 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.276324] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 923.276862] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 923.277184] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Deleting the datastore file [datastore2] 72818e40-624b-4c04-888b-bb622f7f96d7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.277515] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01003394-ba3d-4af3-a6c8-13b0d891d6d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.284086] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce07155-abc7-4c00-b684-8dcdf11eac0f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.291805] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for the task: (returnval){ [ 923.291805] env[61947]: value = "task-1224659" [ 923.291805] env[61947]: _type = "Task" [ 923.291805] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.313031] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.325291] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 97%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.517161] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224652, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.517078} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.517531] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ef8d583a-38fd-4d74-9ed7-f1718e80cace/ef8d583a-38fd-4d74-9ed7-f1718e80cace.vmdk to [datastore1] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 923.518668] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07969b1-d268-4012-9fb5-b4e3e02a8472 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.550887] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.551818] env[61947]: DEBUG oslo_concurrency.lockutils [req-c34a222d-4a2e-4ff5-8ddf-cc0a8cbf5e81 req-810b657a-8630-4373-9119-d19bfbd20a50 service nova] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.552335] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d608e7b3-479e-4ed3-be2a-38233da03226 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.571713] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] Acquired lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.572083] env[61947]: DEBUG nova.network.neutron [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Refreshing network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.584558] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 923.584558] env[61947]: value = "task-1224660" [ 923.584558] env[61947]: _type = "Task" [ 923.584558] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.599559] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.616631] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.428s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.616992] env[61947]: INFO nova.compute.manager [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Migrating [ 923.633850] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.068s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.635105] env[61947]: DEBUG nova.objects.instance [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lazy-loading 'resources' on Instance uuid f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.663716] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.719061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.719061] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Instance network_info: |[{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.719061] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:7a:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b931c4c-f73c-4fbd-9c9f-0270834cc69e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3adc00d1-4b5f-4857-a2ae-bdc4201f8c69', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.728672] env[61947]: DEBUG oslo.service.loopingcall [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.728672] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 923.730148] env[61947]: DEBUG nova.compute.manager [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Received event network-changed-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.730345] env[61947]: DEBUG nova.compute.manager [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Refreshing instance network info cache due to event network-changed-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 923.730570] env[61947]: DEBUG oslo_concurrency.lockutils [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] Acquiring lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.730710] env[61947]: DEBUG oslo_concurrency.lockutils [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] Acquired lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.730911] env[61947]: DEBUG nova.network.neutron [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Refreshing network info cache for port 3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.732714] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3974473-48b3-4831-9927-1f361824b32c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.755377] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.755377] env[61947]: value = "task-1224661" [ 923.755377] env[61947]: _type = "Task" [ 923.755377] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.805720] env[61947]: DEBUG oslo_vmware.api [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Task: {'id': task-1224659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253721} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.806062] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.806600] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.806842] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.807185] env[61947]: INFO nova.compute.manager [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Took 1.17 seconds to destroy the instance on the hypervisor. [ 923.808202] env[61947]: DEBUG oslo.service.loopingcall [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.808375] env[61947]: DEBUG nova.compute.manager [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.808554] env[61947]: DEBUG nova.network.neutron [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 923.821509] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task} progress is 98%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.100574] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.144044] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.144309] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.144806] env[61947]: DEBUG nova.network.neutron [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.162706] env[61947]: DEBUG oslo_vmware.api [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Task: {'id': task-1224656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.029616} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.164223] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.164930] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 924.165482] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 924.165987] env[61947]: INFO nova.compute.manager [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Took 2.18 seconds to destroy the instance on the hypervisor. [ 924.167561] env[61947]: DEBUG oslo.service.loopingcall [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.169105] env[61947]: DEBUG nova.compute.manager [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 924.169105] env[61947]: DEBUG nova.network.neutron [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.271202] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224661, 'name': CreateVM_Task, 'duration_secs': 0.451416} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.274353] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.275529] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.275870] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.276321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.278017] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4109c5ec-8a61-4e57-98bc-1d3545d40e41 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.283363] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 924.283363] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522d03a8-d95f-1031-5357-e60e744b4159" [ 924.283363] env[61947]: _type = "Task" [ 924.283363] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.300420] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522d03a8-d95f-1031-5357-e60e744b4159, 'name': SearchDatastore_Task, 'duration_secs': 0.011465} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.300420] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.300420] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.300420] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.300420] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.300420] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.300420] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8638baee-6118-4da3-8e50-ace22f5714bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.314310] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.314560] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.320853] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31107da7-7999-4086-87ac-b1f24deade70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.323425] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224647, 'name': RelocateVM_Task, 'duration_secs': 3.902679} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.329879] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 924.330192] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264671', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'name': 'volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '91205e54-6bcb-403c-8308-0f0692fe4fba', 'attached_at': '', 'detached_at': '', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'serial': '4639dbb5-6722-408e-92c6-d142b4cdeb3b'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 924.331848] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9686832-a80d-4960-8d99-fcdf1fa2ddc0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.336857] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 924.336857] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bc5deb-ca60-8ea5-7682-4d15290dc9e1" [ 924.336857] env[61947]: _type = "Task" [ 924.336857] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.357144] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abe1266-fc90-4fd2-8fb1-71a55ec4700b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.364334] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bc5deb-ca60-8ea5-7682-4d15290dc9e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009938} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.368464] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff8e4a78-353b-462f-a387-70df43e24c6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.389813] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b/volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.393077] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6b934d0-fc05-4929-8218-2fd93152557d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.409741] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 924.409741] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b8bad7-88ad-e52d-cba4-71e54acae395" [ 924.409741] env[61947]: _type = "Task" [ 924.409741] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.417459] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 924.417459] env[61947]: value = "task-1224662" [ 924.417459] env[61947]: _type = "Task" [ 924.417459] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.421111] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b8bad7-88ad-e52d-cba4-71e54acae395, 'name': SearchDatastore_Task, 'duration_secs': 0.011602} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.424566] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.424708] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.425409] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fd4e06f-c269-423a-9b85-acbd741e819d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.434179] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.435286] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 924.435286] env[61947]: value = "task-1224663" [ 924.435286] env[61947]: _type = "Task" [ 924.435286] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.447713] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.463380] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.463567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.463752] env[61947]: DEBUG nova.network.neutron [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.484198] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3521a3f6-2003-44ed-a8c1-a5c7fab3d35e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.496702] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a71c19-cb3b-42c8-a4f8-eaf38c73bd56 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.535252] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ff5bd7-7e36-445f-bcc1-9589639dfa93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.543992] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d809477-f5e3-4461-aa27-5056d4c09ed4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.559072] env[61947]: DEBUG nova.compute.provider_tree [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.596185] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224660, 'name': ReconfigVM_Task, 'duration_secs': 0.864073} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.597175] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Reconfigured VM instance instance-00000036 to attach disk [datastore1] fa0eed6e-831f-4e7a-9890-fc5a02609aa4/fa0eed6e-831f-4e7a-9890-fc5a02609aa4.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.597339] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecdf312d-8f1c-4503-863a-f730d671479a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.605069] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 924.605069] env[61947]: value = "task-1224664" [ 924.605069] env[61947]: _type = "Task" [ 924.605069] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.615305] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224664, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.633801] env[61947]: DEBUG nova.network.neutron [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updated VIF entry in instance network info cache for port ab2d3df7-884a-42c4-83db-ff3f86592667. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.634306] env[61947]: DEBUG nova.network.neutron [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [{"id": "ab2d3df7-884a-42c4-83db-ff3f86592667", "address": "fa:16:3e:b0:19:40", "network": {"id": "d13ceedc-e2ba-49cb-b150-842102e9da23", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1615131505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "86f19f3f002c4686a26d843a4c819d89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e0c77754-4085-434b-a3e8-d61be099ac67", "external-id": "nsx-vlan-transportzone-822", "segmentation_id": 822, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2d3df7-88", "ovs_interfaceid": "ab2d3df7-884a-42c4-83db-ff3f86592667", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.670575] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Successfully updated port: c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.781105] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 924.781544] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264693', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'name': 'volume-914190f7-1c29-41db-8c24-20092cdc4528', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c342d64-75f6-4b7d-829d-889eb3b6122f', 'attached_at': '', 'detached_at': '', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'serial': '914190f7-1c29-41db-8c24-20092cdc4528'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 924.782546] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87bdef48-c06b-41ec-ac8f-377838ad02f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.802177] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa9481a-af1e-42b2-b977-ce66c324033a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.843673] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] volume-914190f7-1c29-41db-8c24-20092cdc4528/volume-914190f7-1c29-41db-8c24-20092cdc4528.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.844279] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-867f5d14-51b1-46d6-abf1-07d69b8325bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.865719] env[61947]: DEBUG nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-deleted-4e1cb946-7339-4abc-a85d-28eb52415eb9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.865967] env[61947]: INFO nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Neutron deleted interface 4e1cb946-7339-4abc-a85d-28eb52415eb9; detaching it from the instance and deleting it from the info cache [ 924.866296] env[61947]: DEBUG nova.network.neutron [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ef85f03-b199-4718-b501-0b5978b80fe0", "address": "fa:16:3e:c7:85:04", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ef85f03-b1", "ovs_interfaceid": "7ef85f03-b199-4718-b501-0b5978b80fe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.876682] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 924.876682] env[61947]: value = "task-1224665" [ 924.876682] env[61947]: _type = "Task" [ 924.876682] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.890645] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224665, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.935183] env[61947]: DEBUG nova.network.neutron [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updated VIF entry in instance network info cache for port 3adc00d1-4b5f-4857-a2ae-bdc4201f8c69. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.935615] env[61947]: DEBUG nova.network.neutron [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.943834] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224662, 'name': ReconfigVM_Task, 'duration_secs': 0.324706} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.945137] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Reconfigured VM instance instance-0000004c to attach disk [datastore1] volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b/volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.953254] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8f38ba0-7e91-42fc-8d06-32fdf88cb179 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.962424] env[61947]: DEBUG nova.network.neutron [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.972485] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472889} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.974657] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 924.974801] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.975019] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 924.975019] env[61947]: value = "task-1224666" [ 924.975019] env[61947]: _type = "Task" [ 924.975019] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.975814] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e74ba2ed-bee2-419c-abc7-c09a31adaa39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.989620] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224666, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.991151] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 924.991151] env[61947]: value = "task-1224667" [ 924.991151] env[61947]: _type = "Task" [ 924.991151] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.000559] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224667, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.062677] env[61947]: DEBUG nova.scheduler.client.report [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.117574] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224664, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.137069] env[61947]: DEBUG oslo_concurrency.lockutils [req-a9af1f4b-2408-4531-9c60-8df498199414 req-f0f20a2c-562f-482a-94b6-d8829c403ba8 service nova] Releasing lock "refresh_cache-055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.173995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.174132] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.174262] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.260256] env[61947]: DEBUG nova.network.neutron [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.366233] env[61947]: INFO nova.network.neutron [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Port 4e1cb946-7339-4abc-a85d-28eb52415eb9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 925.366356] env[61947]: INFO nova.network.neutron [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Port 7ef85f03-b199-4718-b501-0b5978b80fe0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 925.366667] env[61947]: DEBUG nova.network.neutron [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [{"id": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "address": "fa:16:3e:9f:32:30", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b562a73-9b", "ovs_interfaceid": "1b562a73-9b70-4c31-9d83-f291e5b2cb76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.369798] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.370313] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Acquired lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.371008] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c91bd2a-654e-428b-803a-c6658f2c6695 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.393892] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Releasing lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.394406] env[61947]: WARNING nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Detach interface failed, port_id=4e1cb946-7339-4abc-a85d-28eb52415eb9, reason: No device with interface-id 4e1cb946-7339-4abc-a85d-28eb52415eb9 exists on VM: nova.exception.NotFound: No device with interface-id 4e1cb946-7339-4abc-a85d-28eb52415eb9 exists on VM [ 925.395157] env[61947]: DEBUG nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Received event network-vif-plugged-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.395157] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Acquiring lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.395157] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.395329] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.395474] env[61947]: DEBUG nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] No waiting events found dispatching network-vif-plugged-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.395642] env[61947]: WARNING nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Received unexpected event network-vif-plugged-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 for instance with vm_state building and task_state spawning. [ 925.395813] env[61947]: DEBUG nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Received event network-vif-deleted-c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.395979] env[61947]: INFO nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Neutron deleted interface c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3; detaching it from the instance and deleting it from the info cache [ 925.396176] env[61947]: DEBUG nova.network.neutron [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.404560] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224665, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.442809] env[61947]: DEBUG oslo_concurrency.lockutils [req-ae116c9c-cb8e-4efd-bfa0-5ff4c881cc8f req-ed866e8d-908d-48ea-ac3f-3ec104e32967 service nova] Releasing lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.464471] env[61947]: DEBUG nova.network.neutron [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.470463] env[61947]: INFO nova.compute.manager [-] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Took 1.66 seconds to deallocate network for instance. [ 925.490872] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224666, 'name': ReconfigVM_Task, 'duration_secs': 0.132399} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.491213] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264671', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'name': 'volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '91205e54-6bcb-403c-8308-0f0692fe4fba', 'attached_at': '', 'detached_at': '', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'serial': '4639dbb5-6722-408e-92c6-d142b4cdeb3b'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 925.491799] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8af06eff-8989-4cc2-b344-a705189b3603 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.503901] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224667, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07472} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.505461] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.506244] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 925.506244] env[61947]: value = "task-1224668" [ 925.506244] env[61947]: _type = "Task" [ 925.506244] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.507040] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baa51a0-45e4-4fcb-af4c-266f9832f99a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.525736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.526070] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.526380] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.526606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.526832] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.538039] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.538695] env[61947]: INFO nova.compute.manager [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Terminating instance [ 925.544730] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-505485e1-4c67-442d-ac83-beffe3a15124 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.559526] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224668, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.560299] env[61947]: DEBUG nova.compute.manager [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.560514] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 925.561621] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e6d685-daf6-4f46-a50c-50d4142a3d5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.570972] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.573009] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.574633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.497s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.574959] env[61947]: DEBUG nova.objects.instance [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'resources' on Instance uuid 9cfa7c52-2250-49ce-a1c8-baad6644f824 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.576091] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-976b31cd-e711-4d99-85d6-0136f0967f77 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.578350] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 925.578350] env[61947]: value = "task-1224669" [ 925.578350] env[61947]: _type = "Task" [ 925.578350] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.585741] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 925.585741] env[61947]: value = "task-1224670" [ 925.585741] env[61947]: _type = "Task" [ 925.585741] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.589410] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224669, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.598609] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.600260] env[61947]: INFO nova.scheduler.client.report [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted allocations for instance f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b [ 925.616176] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224664, 'name': Rename_Task, 'duration_secs': 0.936377} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.616522] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 925.616772] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52ca7265-f644-4b7d-b2e8-832c172be0d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.624366] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 925.624366] env[61947]: value = "task-1224671" [ 925.624366] env[61947]: _type = "Task" [ 925.624366] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.633587] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.723010] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.762148] env[61947]: DEBUG nova.compute.manager [req-510ba4ad-6eff-4d69-a0b8-d80d812d86ff req-881cbd35-2e52-46d9-8e3b-a2d4122e2482 service nova] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Received event network-vif-deleted-ab2d3df7-884a-42c4-83db-ff3f86592667 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.763371] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.869837] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.890384] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224665, 'name': ReconfigVM_Task, 'duration_secs': 1.009942} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.891502] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfigured VM instance instance-00000042 to attach disk [datastore2] volume-914190f7-1c29-41db-8c24-20092cdc4528/volume-914190f7-1c29-41db-8c24-20092cdc4528.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.897173] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eba84054-63e3-4121-ba42-4568517775b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.908262] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d922b774-3b08-4f35-9be6-f16065ba2cc2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.919250] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba8de24-854a-4ada-a380-24444964c7e7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.930218] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 925.930218] env[61947]: value = "task-1224672" [ 925.930218] env[61947]: _type = "Task" [ 925.930218] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.931275] env[61947]: DEBUG nova.network.neutron [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updating instance_info_cache with network_info: [{"id": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "address": "fa:16:3e:bc:68:90", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc70b72e2-fd", "ovs_interfaceid": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.942290] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.956206] env[61947]: DEBUG nova.compute.manager [req-2edb4650-8ae3-4c44-8c54-a8cb1cb015eb req-ef642f73-ad98-4b5a-87ba-a97cdb1df8d7 service nova] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Detach interface failed, port_id=c069a41b-ac9b-40ed-a872-ee0f7bbcc4c3, reason: Instance 72818e40-624b-4c04-888b-bb622f7f96d7 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 925.967435] env[61947]: INFO nova.compute.manager [-] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Took 1.80 seconds to deallocate network for instance. [ 925.981021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.021715] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224668, 'name': Rename_Task, 'duration_secs': 0.136947} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.021961] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 926.022242] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7c8a28f-a48b-46db-a671-fb403606bef2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.031341] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 926.031341] env[61947]: value = "task-1224673" [ 926.031341] env[61947]: _type = "Task" [ 926.031341] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.040735] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.079313] env[61947]: DEBUG nova.objects.instance [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'numa_topology' on Instance uuid 9cfa7c52-2250-49ce-a1c8-baad6644f824 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.090778] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224669, 'name': ReconfigVM_Task, 'duration_secs': 0.33603} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.093980] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfigured VM instance instance-0000004d to attach disk [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.095211] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18ac10b0-715f-4d29-b7a3-b613485cc3e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.102398] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224670, 'name': PowerOffVM_Task, 'duration_secs': 0.258014} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.103700] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 926.103884] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 926.104198] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 926.104198] env[61947]: value = "task-1224674" [ 926.104198] env[61947]: _type = "Task" [ 926.104198] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.104426] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd7e99ce-d9f0-4e8c-9d18-22cc7f125946 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.111709] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8773c5e-5753-4905-b7c8-2bc8c0c3853b tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.634s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.119242] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224674, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.135687] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224671, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.229749] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 926.230137] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 926.230416] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleting the datastore file [datastore1] 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.230793] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87b17926-3dba-49d6-abf0-a2b942e357d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.243313] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 926.243313] env[61947]: value = "task-1224676" [ 926.243313] env[61947]: _type = "Task" [ 926.243313] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.258880] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.375989] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e7dbabba-4e29-41c1-94ab-16bc13ae3acd tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-04e5cc55-71d9-4d3d-95c1-fb1401ab74f8-4e1cb946-7339-4abc-a85d-28eb52415eb9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.098s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.434427] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.434855] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Instance network_info: |[{"id": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "address": "fa:16:3e:bc:68:90", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc70b72e2-fd", "ovs_interfaceid": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.444973] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:68:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c70b72e2-fda2-45c5-afe8-1af9c339fdb2', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.452983] env[61947]: DEBUG oslo.service.loopingcall [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.453198] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 926.453926] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29f67f57-bd8e-433e-aa1e-f7e42376c43f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.472646] env[61947]: DEBUG oslo_vmware.api [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224672, 'name': ReconfigVM_Task, 'duration_secs': 0.15845} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.473526] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264693', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'name': 'volume-914190f7-1c29-41db-8c24-20092cdc4528', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c342d64-75f6-4b7d-829d-889eb3b6122f', 'attached_at': '', 'detached_at': '', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'serial': '914190f7-1c29-41db-8c24-20092cdc4528'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 926.475673] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.478953] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.478953] env[61947]: value = "task-1224677" [ 926.478953] env[61947]: _type = "Task" [ 926.478953] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.487378] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224677, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.541801] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224673, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.588448] env[61947]: DEBUG nova.objects.base [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Object Instance<9cfa7c52-2250-49ce-a1c8-baad6644f824> lazy-loaded attributes: resources,numa_topology {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 926.627943] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224674, 'name': Rename_Task, 'duration_secs': 0.167408} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.632543] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 926.636922] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0dbe8f0f-8909-443e-882c-62ab030aad36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.648927] env[61947]: DEBUG oslo_vmware.api [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224671, 'name': PowerOnVM_Task, 'duration_secs': 0.718488} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.651191] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 926.653467] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 926.653467] env[61947]: value = "task-1224678" [ 926.653467] env[61947]: _type = "Task" [ 926.653467] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.666973] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.760066] env[61947]: DEBUG oslo_vmware.api [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229161} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.761193] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.761924] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.762067] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.762262] env[61947]: INFO nova.compute.manager [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Took 1.20 seconds to destroy the instance on the hypervisor. [ 926.762513] env[61947]: DEBUG oslo.service.loopingcall [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.765088] env[61947]: DEBUG nova.compute.manager [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.765194] env[61947]: DEBUG nova.network.neutron [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 926.772944] env[61947]: DEBUG nova.compute.manager [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.773938] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ab165c-832c-4383-88a6-9fbaf9d9e8a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.841329] env[61947]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 7ef85f03-b199-4718-b501-0b5978b80fe0 could not be found.", "detail": ""}} {{(pid=61947) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 926.841590] env[61947]: DEBUG nova.network.neutron [-] Unable to show port 7ef85f03-b199-4718-b501-0b5978b80fe0 as it no longer exists. {{(pid=61947) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 926.857387] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24c01e1-388b-4908-9478-a92daf35bf85 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.866680] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31215db-a270-4bc9-8eee-ee93543296ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.912294] env[61947]: DEBUG nova.compute.manager [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Received event network-changed-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.912529] env[61947]: DEBUG nova.compute.manager [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Refreshing instance network info cache due to event network-changed-c70b72e2-fda2-45c5-afe8-1af9c339fdb2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.912709] env[61947]: DEBUG oslo_concurrency.lockutils [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] Acquiring lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.912861] env[61947]: DEBUG oslo_concurrency.lockutils [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] Acquired lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.913280] env[61947]: DEBUG nova.network.neutron [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Refreshing network info cache for port c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.915051] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769338c1-1051-48e1-9fb1-20485bf37398 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.926108] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fee430a-80be-42ba-8499-abd6a3c61585 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.948571] env[61947]: DEBUG nova.compute.provider_tree [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.990050] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224677, 'name': CreateVM_Task, 'duration_secs': 0.498812} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.990372] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 926.991660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.991933] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.992341] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.993037] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d2efe7d-aaaf-4c18-b4eb-9010a44c8496 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.998844] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 926.998844] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]529c1b14-5ab3-d1e1-6bbf-771a37d1f49f" [ 926.998844] env[61947]: _type = "Task" [ 926.998844] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.010042] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]529c1b14-5ab3-d1e1-6bbf-771a37d1f49f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.047073] env[61947]: DEBUG oslo_vmware.api [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224673, 'name': PowerOnVM_Task, 'duration_secs': 0.546978} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.047073] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.047073] env[61947]: INFO nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Took 8.61 seconds to spawn the instance on the hypervisor. [ 927.047073] env[61947]: DEBUG nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.047073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f642f8-b2ba-4820-9e7e-46ab91f71a4a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.119048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.121603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.121603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.121603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.121603] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.122763] env[61947]: INFO nova.compute.manager [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Terminating instance [ 927.124934] env[61947]: DEBUG nova.compute.manager [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.126130] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 927.126624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ef87c2-b315-4ef3-a8ac-db2c6ebe0a84 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.137206] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.137694] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38971d11-31a6-4e68-a7b6-caa042d2027c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.149024] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 927.149024] env[61947]: value = "task-1224679" [ 927.149024] env[61947]: _type = "Task" [ 927.149024] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.161998] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.169507] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224678, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.300204] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb53c76-e6e8-44a6-9f98-485cddd48ea5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.306806] env[61947]: DEBUG oslo_concurrency.lockutils [None req-339885d4-4a7d-40fc-ba44-de4f5bf8c63d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.827s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.327313] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 927.454740] env[61947]: DEBUG nova.scheduler.client.report [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.510844] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]529c1b14-5ab3-d1e1-6bbf-771a37d1f49f, 'name': SearchDatastore_Task, 'duration_secs': 0.017399} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.511371] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.511498] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.511720] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.511859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.511966] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.512252] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13865f2e-f10e-431a-baa4-a7b24f41db3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.519682] env[61947]: DEBUG nova.objects.instance [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lazy-loading 'flavor' on Instance uuid 8c342d64-75f6-4b7d-829d-889eb3b6122f {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.527925] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.528251] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 927.528965] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f384629-6f4a-4374-bbde-f8f6a26b6167 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.535980] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 927.535980] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52de114f-fac7-2dc6-2a87-260d712f4f06" [ 927.535980] env[61947]: _type = "Task" [ 927.535980] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.549305] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52de114f-fac7-2dc6-2a87-260d712f4f06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.566626] env[61947]: INFO nova.compute.manager [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Took 28.30 seconds to build instance. [ 927.661351] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224679, 'name': PowerOffVM_Task, 'duration_secs': 0.338881} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.666137] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.666494] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.666836] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-341c7e13-f5d9-4311-a2ab-762929c90fce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.678311] env[61947]: DEBUG oslo_vmware.api [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224678, 'name': PowerOnVM_Task, 'duration_secs': 0.736634} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.681018] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.681018] env[61947]: INFO nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Took 6.80 seconds to spawn the instance on the hypervisor. [ 927.681157] env[61947]: DEBUG nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.682165] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f22a9e-aca6-4188-afc6-6f0f2ee91883 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.759850] env[61947]: DEBUG nova.network.neutron [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updated VIF entry in instance network info cache for port c70b72e2-fda2-45c5-afe8-1af9c339fdb2. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.760255] env[61947]: DEBUG nova.network.neutron [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updating instance_info_cache with network_info: [{"id": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "address": "fa:16:3e:bc:68:90", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc70b72e2-fd", "ovs_interfaceid": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.776367] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.776810] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.776922] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleting the datastore file [datastore2] 9bc18645-c837-4dab-8e05-fbcf0d40812f {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.777140] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ca24dce-d501-4a0a-81d1-4c8e01121934 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.787023] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 927.787023] env[61947]: value = "task-1224681" [ 927.787023] env[61947]: _type = "Task" [ 927.787023] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.797149] env[61947]: DEBUG nova.compute.manager [req-7f00d95f-cb76-4905-8fa6-b46fe1196a03 req-2ee4e423-56c6-455c-9045-d1016bc5a069 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-deleted-1b562a73-9b70-4c31-9d83-f291e5b2cb76 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.797503] env[61947]: INFO nova.compute.manager [req-7f00d95f-cb76-4905-8fa6-b46fe1196a03 req-2ee4e423-56c6-455c-9045-d1016bc5a069 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Neutron deleted interface 1b562a73-9b70-4c31-9d83-f291e5b2cb76; detaching it from the instance and deleting it from the info cache [ 927.797503] env[61947]: DEBUG nova.network.neutron [req-7f00d95f-cb76-4905-8fa6-b46fe1196a03 req-2ee4e423-56c6-455c-9045-d1016bc5a069 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.808564] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.835135] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.837017] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0133e9d-3cc0-4dea-a9ca-8b04e70b38a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.844281] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 927.844281] env[61947]: value = "task-1224682" [ 927.844281] env[61947]: _type = "Task" [ 927.844281] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.861657] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 927.861895] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 927.871038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "e6435156-fb2e-4912-9587-99eeafeded87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.871176] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.966725] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.390s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.967162] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.990s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.967393] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.970048] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.494s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.970393] env[61947]: DEBUG nova.objects.instance [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lazy-loading 'resources' on Instance uuid 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.996551] env[61947]: INFO nova.scheduler.client.report [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Deleted allocations for instance 72818e40-624b-4c04-888b-bb622f7f96d7 [ 927.999848] env[61947]: WARNING oslo_messaging._drivers.amqpdriver [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 928.029423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1caebe12-e0e1-4043-b2f6-9bc02273ef6c tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.927s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.053166] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52de114f-fac7-2dc6-2a87-260d712f4f06, 'name': SearchDatastore_Task, 'duration_secs': 0.015721} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.053166] env[61947]: DEBUG nova.network.neutron [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.053166] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4517b6eb-ae89-40b0-aea0-5aacd9692d16 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.063410] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 928.063410] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52835330-b8bd-50a1-a3af-b8dbfea77f03" [ 928.063410] env[61947]: _type = "Task" [ 928.063410] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.076607] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a83a74aa-6863-4d05-acb2-52357e7ceb5f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.819s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.077284] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52835330-b8bd-50a1-a3af-b8dbfea77f03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.205573] env[61947]: INFO nova.compute.manager [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Took 17.71 seconds to build instance. [ 928.264475] env[61947]: DEBUG oslo_concurrency.lockutils [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] Releasing lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.264587] env[61947]: DEBUG nova.compute.manager [req-85e1a9ed-6cc9-4d7a-9298-be867c7c764f req-cff023ba-5603-41ee-ad96-e14c2c8b30e9 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Received event network-vif-deleted-7ef85f03-b199-4718-b501-0b5978b80fe0 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.298475] env[61947]: DEBUG oslo_vmware.api [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314456} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.298655] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.298839] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 928.299040] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.299224] env[61947]: INFO nova.compute.manager [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 928.299558] env[61947]: DEBUG oslo.service.loopingcall [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.299742] env[61947]: DEBUG nova.compute.manager [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.299838] env[61947]: DEBUG nova.network.neutron [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.302425] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f65d223d-ebc6-4f18-ba4c-4d7ee2909eff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.312626] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8318ec-1d6c-4161-aa57-a59d03202041 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.346500] env[61947]: DEBUG nova.compute.manager [req-7f00d95f-cb76-4905-8fa6-b46fe1196a03 req-2ee4e423-56c6-455c-9045-d1016bc5a069 service nova] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Detach interface failed, port_id=1b562a73-9b70-4c31-9d83-f291e5b2cb76, reason: Instance 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 928.376317] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.376610] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.376797] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.376993] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.377163] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.377315] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.377525] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.377691] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.377868] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.378121] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.378342] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.383492] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.387530] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-899a314d-047d-4a9d-bf30-28133647c513 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.406766] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 928.406766] env[61947]: value = "task-1224683" [ 928.406766] env[61947]: _type = "Task" [ 928.406766] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.417482] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224683, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.479466] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5306891a-64dd-4d3f-be23-2ae90296e00d tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.841s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.480560] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 7.484s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.480798] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.481032] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.481250] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.483682] env[61947]: INFO nova.compute.manager [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Terminating instance [ 928.488324] env[61947]: DEBUG nova.compute.manager [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.488529] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 928.488847] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc6074ff-bac3-413c-aa46-2ce929260f88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.500661] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc5ea41-e0e4-4c86-9881-c95752138ce0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.521039] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2fea6ebc-973c-46a1-bcdb-df192c433e43 tempest-MigrationsAdminTest-1711979822 tempest-MigrationsAdminTest-1711979822-project-member] Lock "72818e40-624b-4c04-888b-bb622f7f96d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.896s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.545797] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9cfa7c52-2250-49ce-a1c8-baad6644f824 could not be found. [ 928.545797] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.546056] env[61947]: INFO nova.compute.manager [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Took 0.06 seconds to destroy the instance on the hypervisor. [ 928.546314] env[61947]: DEBUG oslo.service.loopingcall [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.550238] env[61947]: DEBUG nova.compute.manager [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.550238] env[61947]: DEBUG nova.network.neutron [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.558253] env[61947]: INFO nova.compute.manager [-] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Took 1.79 seconds to deallocate network for instance. [ 928.582035] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52835330-b8bd-50a1-a3af-b8dbfea77f03, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.582035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.582239] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] c4647324-7e52-44d4-a4b0-961684cc1c4c/c4647324-7e52-44d4-a4b0-961684cc1c4c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 928.582618] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f4723e8-91d0-48e3-95f9-e7cc0573b59d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.593628] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 928.593628] env[61947]: value = "task-1224684" [ 928.593628] env[61947]: _type = "Task" [ 928.593628] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.603234] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.710209] env[61947]: DEBUG oslo_concurrency.lockutils [None req-477453b8-5302-47a5-aada-3cad747614e6 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.222s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.799329] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed50bf62-1281-4303-bb69-aa2fd591650c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.808801] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb23b5d-57e2-4e3e-a5a3-8e9c3c0709af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.856347] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d916104d-ec44-4e2d-b574-a5150c1c5f5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.866731] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a2745b-c50a-4b77-b8be-a04d06468f8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.883818] env[61947]: DEBUG nova.compute.provider_tree [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.920870] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224683, 'name': ReconfigVM_Task, 'duration_secs': 0.280858} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.923753] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.923753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 929.025165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.025165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.068113] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.107151] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224684, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.209455] env[61947]: DEBUG nova.compute.manager [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Received event network-changed-c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.209934] env[61947]: DEBUG nova.compute.manager [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Refreshing instance network info cache due to event network-changed-c2f0d861-74c8-4b27-abb4-94bfa51b7768. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 929.210326] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] Acquiring lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.210479] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] Acquired lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.210737] env[61947]: DEBUG nova.network.neutron [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Refreshing network info cache for port c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.233930] env[61947]: DEBUG nova.network.neutron [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.310017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.310243] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.388246] env[61947]: DEBUG nova.scheduler.client.report [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.430250] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.430687] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.430973] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.431287] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.431530] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.431785] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.432324] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.432391] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.432638] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.432902] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.433214] env[61947]: DEBUG nova.virt.hardware [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.439802] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfiguring VM instance instance-00000034 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 929.439933] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8af19693-e7e7-4f82-b3a5-686a29833f4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.454729] env[61947]: DEBUG nova.network.neutron [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.460609] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 929.460609] env[61947]: value = "task-1224685" [ 929.460609] env[61947]: _type = "Task" [ 929.460609] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.470609] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.530863] env[61947]: INFO nova.compute.manager [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Detaching volume 914190f7-1c29-41db-8c24-20092cdc4528 [ 929.575616] env[61947]: INFO nova.virt.block_device [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Attempting to driver detach volume 914190f7-1c29-41db-8c24-20092cdc4528 from mountpoint /dev/sdb [ 929.575883] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 929.576081] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264693', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'name': 'volume-914190f7-1c29-41db-8c24-20092cdc4528', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c342d64-75f6-4b7d-829d-889eb3b6122f', 'attached_at': '', 'detached_at': '', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'serial': '914190f7-1c29-41db-8c24-20092cdc4528'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 929.576957] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdbef97-0c09-40ca-9ff1-d0c786a7a89f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.602302] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d14601-6f67-4bcb-aceb-ba66052dfa64 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.610841] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603069} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.612641] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] c4647324-7e52-44d4-a4b0-961684cc1c4c/c4647324-7e52-44d4-a4b0-961684cc1c4c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 929.612883] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.613266] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd12ab38-7ce2-43f6-9167-786151766554 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.615473] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670aa75c-ee07-4d90-8889-3d50a2587cdf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.637873] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882a4d1d-4af9-4d6e-8d7c-d20f191a945f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.641051] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 929.641051] env[61947]: value = "task-1224686" [ 929.641051] env[61947]: _type = "Task" [ 929.641051] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.656727] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] The volume has not been displaced from its original location: [datastore2] volume-914190f7-1c29-41db-8c24-20092cdc4528/volume-914190f7-1c29-41db-8c24-20092cdc4528.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 929.662600] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 929.666651] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cf19c9f-e74f-4b7a-90dd-5629ecc6d109 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.683268] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.688979] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 929.688979] env[61947]: value = "task-1224687" [ 929.688979] env[61947]: _type = "Task" [ 929.688979] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.700768] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.737988] env[61947]: INFO nova.compute.manager [-] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Took 1.44 seconds to deallocate network for instance. [ 929.745478] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.745786] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.812525] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 929.900570] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.904891] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.981s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.907189] env[61947]: INFO nova.compute.claims [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.936061] env[61947]: DEBUG nova.compute.manager [req-318adefd-e855-4c6b-981f-84f065fdae46 req-47780d7a-bcec-48b2-a0e1-96e216570918 service nova] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Received event network-vif-deleted-8fa0162f-f4dc-4407-9a88-0f6c75ffd12a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.936993] env[61947]: INFO nova.scheduler.client.report [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Deleted allocations for instance 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb [ 929.962190] env[61947]: INFO nova.compute.manager [-] [instance: 9cfa7c52-2250-49ce-a1c8-baad6644f824] Took 1.41 seconds to deallocate network for instance. [ 929.982737] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224685, 'name': ReconfigVM_Task, 'duration_secs': 0.33658} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.982737] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfigured VM instance instance-00000034 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 929.982737] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86652fd3-2e62-4d34-826f-d5d34369c8a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.009387] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.010865] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b1bef42-42e9-4697-9547-5cd283cf3b68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.035907] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 930.035907] env[61947]: value = "task-1224688" [ 930.035907] env[61947]: _type = "Task" [ 930.035907] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.044916] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.104766] env[61947]: DEBUG nova.network.neutron [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updated VIF entry in instance network info cache for port c2f0d861-74c8-4b27-abb4-94bfa51b7768. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.105233] env[61947]: DEBUG nova.network.neutron [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updating instance_info_cache with network_info: [{"id": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "address": "fa:16:3e:ae:bb:b9", "network": {"id": "347593d1-862d-4d28-b3ec-328cf05d68f9", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1066216228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b54acf7e242414ab95acfe5433ab171", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2f0d861-74", "ovs_interfaceid": "c2f0d861-74c8-4b27-abb4-94bfa51b7768", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.155203] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069502} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.155941] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.156546] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76220a2f-0cb4-4df7-9302-1f7f0cf69f84 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.182693] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] c4647324-7e52-44d4-a4b0-961684cc1c4c/c4647324-7e52-44d4-a4b0-961684cc1c4c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.183065] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d13f0d9-69df-423e-9b81-63009cfe11fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.210026] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224687, 'name': ReconfigVM_Task, 'duration_secs': 0.34454} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.211647] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 930.216668] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 930.216668] env[61947]: value = "task-1224689" [ 930.216668] env[61947]: _type = "Task" [ 930.216668] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.216954] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07690e31-be80-46c7-81ae-572268805d10 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.240546] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224689, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.242157] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 930.242157] env[61947]: value = "task-1224690" [ 930.242157] env[61947]: _type = "Task" [ 930.242157] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.247137] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.258201] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.258201] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 930.259198] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224690, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.338272] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.454623] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6438aea8-9900-4982-a255-91b2abdcdfbe tempest-ServerRescueTestJSONUnderV235-1366049666 tempest-ServerRescueTestJSONUnderV235-1366049666-project-member] Lock "055a0b7f-1703-447c-b9d4-0a8efc4f5dbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.479s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.547710] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224688, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.608290] env[61947]: DEBUG oslo_concurrency.lockutils [req-bc03dae6-50b7-40fb-b86e-8f18dd8b0651 req-ad0de617-3f9a-4c0a-be7e-b7c886921723 service nova] Releasing lock "refresh_cache-91205e54-6bcb-403c-8308-0f0692fe4fba" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.739367] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224689, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.754171] env[61947]: DEBUG oslo_vmware.api [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224690, 'name': ReconfigVM_Task, 'duration_secs': 0.163616} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.754517] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264693', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'name': 'volume-914190f7-1c29-41db-8c24-20092cdc4528', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c342d64-75f6-4b7d-829d-889eb3b6122f', 'attached_at': '', 'detached_at': '', 'volume_id': '914190f7-1c29-41db-8c24-20092cdc4528', 'serial': '914190f7-1c29-41db-8c24-20092cdc4528'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 930.996601] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a338edfb-807e-4fdb-90d1-d6fbe322930a tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "9cfa7c52-2250-49ce-a1c8-baad6644f824" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.516s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.049515] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224688, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.147391] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063615e5-1334-44b7-a4d2-c65fbeda7dc4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.159391] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99968c0-1910-4f5b-b72b-2a2c83037e51 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.163714] env[61947]: DEBUG nova.compute.manager [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Stashing vm_state: active {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 931.208664] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6e0d86-ea75-412f-9361-8b01bb1bed6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.218076] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d399b5d-b7c9-49fb-a73d-411801dfa1d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.235186] env[61947]: DEBUG nova.compute.provider_tree [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.248154] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224689, 'name': ReconfigVM_Task, 'duration_secs': 0.977984} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.248154] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Reconfigured VM instance instance-0000004e to attach disk [datastore2] c4647324-7e52-44d4-a4b0-961684cc1c4c/c4647324-7e52-44d4-a4b0-961684cc1c4c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.248719] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d24ae47-9a8f-425a-87ac-2066d370576e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.257069] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 931.257069] env[61947]: value = "task-1224691" [ 931.257069] env[61947]: _type = "Task" [ 931.257069] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.272142] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224691, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.312517] env[61947]: DEBUG nova.objects.instance [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lazy-loading 'flavor' on Instance uuid 8c342d64-75f6-4b7d-829d-889eb3b6122f {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.548107] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224688, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.726711] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.742114] env[61947]: DEBUG nova.scheduler.client.report [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.770311] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224691, 'name': Rename_Task, 'duration_secs': 0.300609} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.770548] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 931.770838] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31b480b4-0c68-4512-ba22-712450d568dc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.784117] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 931.784117] env[61947]: value = "task-1224692" [ 931.784117] env[61947]: _type = "Task" [ 931.784117] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.793392] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.048410] env[61947]: DEBUG oslo_vmware.api [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224688, 'name': ReconfigVM_Task, 'duration_secs': 1.588626} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.048764] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Reconfigured VM instance instance-00000034 to attach disk [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657/ab027a9d-f51d-4cb6-8aaf-ebd507548657.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.049075] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 932.248260] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.248978] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.254043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.187s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.254300] env[61947]: DEBUG nova.objects.instance [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'resources' on Instance uuid 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.265089] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "ab21547e-0ddb-4271-967f-bea5056c163a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.265372] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.298359] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224692, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.320849] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc3ea5cd-aad2-4c7f-ab7f-681b954ba43e tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.297s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.439691] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1145b81b-c00b-4244-ae58-da575f64346a tempest-ServersListShow296Test-1724047034 tempest-ServersListShow296Test-1724047034-project-member] Acquiring lock "200019ef-1600-4c51-8b01-aef9127bbc80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.441334] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1145b81b-c00b-4244-ae58-da575f64346a tempest-ServersListShow296Test-1724047034 tempest-ServersListShow296Test-1724047034-project-member] Lock "200019ef-1600-4c51-8b01-aef9127bbc80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.563477] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91ce75f-da76-4604-bdb2-bacf72a837c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.605661] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e682379-6c3a-4c9d-b3b1-87ec5497fba3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.629971] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 932.757641] env[61947]: DEBUG nova.compute.utils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.763516] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.763516] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.767367] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.796449] env[61947]: DEBUG oslo_vmware.api [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224692, 'name': PowerOnVM_Task, 'duration_secs': 0.762113} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.799276] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 932.799514] env[61947]: INFO nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Took 9.57 seconds to spawn the instance on the hypervisor. [ 932.799752] env[61947]: DEBUG nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.800836] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5325b8cf-1e32-4eb0-bad3-c50b9a1c3419 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.828252] env[61947]: DEBUG nova.policy [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41d4a2d39ebc4ce0a0e76350f907e7df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58b83ea896bd4654aaa9e81f635a6a55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 932.944796] env[61947]: DEBUG nova.compute.manager [None req-1145b81b-c00b-4244-ae58-da575f64346a tempest-ServersListShow296Test-1724047034 tempest-ServersListShow296Test-1724047034-project-member] [instance: 200019ef-1600-4c51-8b01-aef9127bbc80] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.999370] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d41e7d7-f4d8-4114-b109-0394b9e2a45d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.007850] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2b112a-570e-4c66-98da-8350563aff19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.040712] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a34f1fc-1118-46b3-a813-64c0c64eae89 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.049691] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fe29ae-7195-4610-82f4-b4336a79cc9c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.067311] env[61947]: DEBUG nova.compute.provider_tree [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.182082] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Successfully created port: 2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.224015] env[61947]: DEBUG nova.network.neutron [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Port 46083345-81bb-4fe1-ac51-a4da371e4a16 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 933.263183] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.287799] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.323700] env[61947]: INFO nova.compute.manager [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Took 20.40 seconds to build instance. [ 933.451582] env[61947]: DEBUG nova.compute.manager [None req-1145b81b-c00b-4244-ae58-da575f64346a tempest-ServersListShow296Test-1724047034 tempest-ServersListShow296Test-1724047034-project-member] [instance: 200019ef-1600-4c51-8b01-aef9127bbc80] Instance disappeared before build. {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 933.570409] env[61947]: DEBUG nova.scheduler.client.report [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.653146] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.653146] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.825772] env[61947]: DEBUG oslo_concurrency.lockutils [None req-857cb55a-b255-4bea-b725-3b644f8d7640 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.913s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.968112] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1145b81b-c00b-4244-ae58-da575f64346a tempest-ServersListShow296Test-1724047034 tempest-ServersListShow296Test-1724047034-project-member] Lock "200019ef-1600-4c51-8b01-aef9127bbc80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.527s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.076153] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.079410] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.833s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.079770] env[61947]: DEBUG nova.objects.instance [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lazy-loading 'resources' on Instance uuid 9bc18645-c837-4dab-8e05-fbcf0d40812f {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.111578] env[61947]: INFO nova.scheduler.client.report [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted allocations for instance 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8 [ 934.154721] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.250077] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.250350] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.250530] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.272491] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.281982] env[61947]: DEBUG nova.compute.manager [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Received event network-changed-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.282221] env[61947]: DEBUG nova.compute.manager [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Refreshing instance network info cache due to event network-changed-c70b72e2-fda2-45c5-afe8-1af9c339fdb2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 934.283145] env[61947]: DEBUG oslo_concurrency.lockutils [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] Acquiring lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.283145] env[61947]: DEBUG oslo_concurrency.lockutils [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] Acquired lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.283145] env[61947]: DEBUG nova.network.neutron [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Refreshing network info cache for port c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.286570] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.286705] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.286859] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 934.304565] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.304825] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.304989] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.305195] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.305346] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.305499] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.305860] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.306155] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.306458] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.306759] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.307058] env[61947]: DEBUG nova.virt.hardware [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.308975] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dca0180-a735-4c4f-a906-a3e480bea1f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.321051] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4560dc0-d45c-4a3e-b2bb-ed27ed6e8eac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.621357] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b4668b5a-a9ac-4e6f-9807-168ec6c3c619 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "04e5cc55-71d9-4d3d-95c1-fb1401ab74f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.095s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.676310] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.852015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4751a9f1-2be5-4ec4-a445-1ee591a3aafc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.868371] env[61947]: DEBUG nova.compute.manager [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Received event network-vif-plugged-2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.869125] env[61947]: DEBUG oslo_concurrency.lockutils [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] Acquiring lock "e6435156-fb2e-4912-9587-99eeafeded87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.869515] env[61947]: DEBUG oslo_concurrency.lockutils [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] Lock "e6435156-fb2e-4912-9587-99eeafeded87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.870046] env[61947]: DEBUG oslo_concurrency.lockutils [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] Lock "e6435156-fb2e-4912-9587-99eeafeded87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.870359] env[61947]: DEBUG nova.compute.manager [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] No waiting events found dispatching network-vif-plugged-2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.870815] env[61947]: WARNING nova.compute.manager [req-e1bf3f9e-f736-42a5-a782-378fa61ddf30 req-71adc784-223b-4b52-a471-7d55f46b26b3 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Received unexpected event network-vif-plugged-2e329412-3c2f-495d-9ed0-0329e52e1c7f for instance with vm_state building and task_state spawning. [ 934.873689] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8815258f-51d5-4a79-ab21-e16a113b78d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.913431] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6ea581-7fdf-46b0-876d-bc8e5b9d8e8e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.924261] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14699164-fac0-4e73-94bc-840e0418d913 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.947443] env[61947]: DEBUG nova.compute.provider_tree [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.133869] env[61947]: DEBUG nova.network.neutron [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updated VIF entry in instance network info cache for port c70b72e2-fda2-45c5-afe8-1af9c339fdb2. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.134229] env[61947]: DEBUG nova.network.neutron [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updating instance_info_cache with network_info: [{"id": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "address": "fa:16:3e:bc:68:90", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc70b72e2-fd", "ovs_interfaceid": "c70b72e2-fda2-45c5-afe8-1af9c339fdb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.297939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.457438] env[61947]: DEBUG nova.scheduler.client.report [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.493441] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Successfully updated port: 2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.527007] env[61947]: DEBUG nova.compute.manager [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Received event network-changed-2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.527992] env[61947]: DEBUG nova.compute.manager [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Refreshing instance network info cache due to event network-changed-2e329412-3c2f-495d-9ed0-0329e52e1c7f. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.527992] env[61947]: DEBUG oslo_concurrency.lockutils [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] Acquiring lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.527992] env[61947]: DEBUG oslo_concurrency.lockutils [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] Acquired lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.527992] env[61947]: DEBUG nova.network.neutron [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Refreshing network info cache for port 2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.609257] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.639758] env[61947]: DEBUG oslo_concurrency.lockutils [req-6595761e-faf2-4cfd-a693-9548889abc1f req-982d175d-3f0e-4e3e-8cfa-459dc52f3592 service nova] Releasing lock "refresh_cache-c4647324-7e52-44d4-a4b0-961684cc1c4c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.964926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.967363] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.629s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.968943] env[61947]: INFO nova.compute.claims [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.986596] env[61947]: INFO nova.scheduler.client.report [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted allocations for instance 9bc18645-c837-4dab-8e05-fbcf0d40812f [ 936.002189] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.068186] env[61947]: DEBUG nova.network.neutron [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.112049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.112235] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 936.112475] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.112661] env[61947]: DEBUG nova.network.neutron [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.113782] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.114400] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.114489] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.114650] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.114846] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.114997] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.115140] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 936.115289] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 936.192885] env[61947]: DEBUG nova.network.neutron [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.494909] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2b32cd89-3813-461f-beef-a526a6dc3d65 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "9bc18645-c837-4dab-8e05-fbcf0d40812f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.375s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.617896] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.695254] env[61947]: DEBUG oslo_concurrency.lockutils [req-17cfc01c-aacd-4999-a9de-63274831f797 req-ab1a4910-6a87-4603-8f47-83a5694b90a0 service nova] Releasing lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.695598] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.695809] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.863389] env[61947]: DEBUG nova.network.neutron [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.005843] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.006135] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.158812] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e79373a-1878-4f71-b7f6-100d548882ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.167423] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf32c0a6-03a5-42cc-b61c-38fdbf57d9db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.196728] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e899486e-bc53-4fe9-a8a8-2246676df34d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.206104] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4484a444-33cb-41a1-83b8-936b037e0d04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.223144] env[61947]: DEBUG nova.compute.provider_tree [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.242146] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.366308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.392047] env[61947]: DEBUG nova.network.neutron [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Updating instance_info_cache with network_info: [{"id": "2e329412-3c2f-495d-9ed0-0329e52e1c7f", "address": "fa:16:3e:fa:22:ef", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e329412-3c", "ovs_interfaceid": "2e329412-3c2f-495d-9ed0-0329e52e1c7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.508164] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.726348] env[61947]: DEBUG nova.scheduler.client.report [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.887604] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d463709-5343-4ab9-a91e-38d7a2aace5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.907360] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "refresh_cache-e6435156-fb2e-4912-9587-99eeafeded87" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.907675] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance network_info: |[{"id": "2e329412-3c2f-495d-9ed0-0329e52e1c7f", "address": "fa:16:3e:fa:22:ef", "network": {"id": "c5c2cef0-de8f-4ea7-afd4-04039757081f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2132766174-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58b83ea896bd4654aaa9e81f635a6a55", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e329412-3c", "ovs_interfaceid": "2e329412-3c2f-495d-9ed0-0329e52e1c7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.908304] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:22:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e329412-3c2f-495d-9ed0-0329e52e1c7f', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.915630] env[61947]: DEBUG oslo.service.loopingcall [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.916320] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8403ac89-7520-4d56-acfb-4ec29f5f4660 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.919091] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 937.919321] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a921d1d4-af1b-457d-9225-d7ff06179e1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.939833] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 937.944607] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.944607] env[61947]: value = "task-1224693" [ 937.944607] env[61947]: _type = "Task" [ 937.944607] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.953481] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224693, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.027373] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.233020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.233603] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.236558] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.510s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.447013] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2217a882-1195-4a94-bd90-16d59ef60dd7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance 'ab027a9d-f51d-4cb6-8aaf-ebd507548657' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 938.462061] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224693, 'name': CreateVM_Task, 'duration_secs': 0.320159} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.462230] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 938.462907] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.463098] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.463426] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.463677] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f732256-29f6-4f1c-a3a5-a6d24e63bb3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.469244] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 938.469244] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520c2b0f-2ecb-b3e4-588d-4a94b5ce8ae0" [ 938.469244] env[61947]: _type = "Task" [ 938.469244] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.477585] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520c2b0f-2ecb-b3e4-588d-4a94b5ce8ae0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.740242] env[61947]: DEBUG nova.compute.utils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.744538] env[61947]: INFO nova.compute.claims [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.748510] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.748719] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.787172] env[61947]: DEBUG nova.policy [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c72bd98842ef47b0a00a8e74cbd02ce9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2f48256eb99494d942f3bb8c4a62a67', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 938.983404] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520c2b0f-2ecb-b3e4-588d-4a94b5ce8ae0, 'name': SearchDatastore_Task, 'duration_secs': 0.011026} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.983780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.984112] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.984934] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.984934] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.984934] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.985104] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd0130e0-d37b-4a3a-a80a-e32221143079 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.994614] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.994765] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 938.995578] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d9cdeac-7471-4a01-b6d8-47275477b263 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.001666] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 939.001666] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cb1b3f-b171-a12f-0bc2-ecab6f95506a" [ 939.001666] env[61947]: _type = "Task" [ 939.001666] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.010297] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cb1b3f-b171-a12f-0bc2-ecab6f95506a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.040631] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Successfully created port: fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.252018] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.253752] env[61947]: INFO nova.compute.resource_tracker [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating resource usage from migration 6a038d9b-8359-4af1-a990-eab5cb334626 [ 939.461272] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdddc784-c893-45d2-b665-32227122937f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.474455] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5589547b-43a8-469b-aa2f-ddac699ffff6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.511094] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402ff9fe-8bec-4463-a5a0-da8ec8a4ad18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.519563] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cb1b3f-b171-a12f-0bc2-ecab6f95506a, 'name': SearchDatastore_Task, 'duration_secs': 0.009218} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.522214] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7e5a015-5aa1-4bd0-9fa3-f4fda9df9292 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.525145] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c063b0-486f-4c40-a48d-c78520ff43db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.540469] env[61947]: DEBUG nova.compute.provider_tree [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.542546] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 939.542546] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527a0c9d-1ca1-e5a7-5e28-0d69d46c0f69" [ 939.542546] env[61947]: _type = "Task" [ 939.542546] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.550657] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527a0c9d-1ca1-e5a7-5e28-0d69d46c0f69, 'name': SearchDatastore_Task, 'duration_secs': 0.00942} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.551463] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.551763] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e6435156-fb2e-4912-9587-99eeafeded87/e6435156-fb2e-4912-9587-99eeafeded87.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 939.552027] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69c95a59-98ea-4a31-87f0-f9dc1432a12e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.560788] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 939.560788] env[61947]: value = "task-1224694" [ 939.560788] env[61947]: _type = "Task" [ 939.560788] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.571787] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.043945] env[61947]: DEBUG nova.scheduler.client.report [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.072209] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224694, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509385} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.072490] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] e6435156-fb2e-4912-9587-99eeafeded87/e6435156-fb2e-4912-9587-99eeafeded87.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.072706] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.072967] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f965822c-9339-444e-8353-c62361ffa24c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.080113] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 940.080113] env[61947]: value = "task-1224695" [ 940.080113] env[61947]: _type = "Task" [ 940.080113] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.090336] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224695, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.263502] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.288082] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.288362] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.288525] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.288738] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.288905] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.289088] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.289307] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.289472] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.289646] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.289813] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.289991] env[61947]: DEBUG nova.virt.hardware [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.290843] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124dbc26-6fff-44af-ad43-c058779d58c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.299514] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72cb33a-af43-4897-894c-28a56552391a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.549226] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.312s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.549422] env[61947]: INFO nova.compute.manager [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Migrating [ 940.556995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.269s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.559059] env[61947]: INFO nova.compute.claims [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.580502] env[61947]: DEBUG nova.compute.manager [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Received event network-vif-plugged-fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.580741] env[61947]: DEBUG oslo_concurrency.lockutils [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] Acquiring lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.580980] env[61947]: DEBUG oslo_concurrency.lockutils [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.581181] env[61947]: DEBUG oslo_concurrency.lockutils [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.581361] env[61947]: DEBUG nova.compute.manager [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] No waiting events found dispatching network-vif-plugged-fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.581537] env[61947]: WARNING nova.compute.manager [req-1dbc884a-8942-4acf-b30c-f084b9421f00 req-841e2c0d-9c2f-4025-ad48-3654da38dac9 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Received unexpected event network-vif-plugged-fe941a11-be7a-47ce-b314-cbeab2ec8cd1 for instance with vm_state building and task_state spawning. [ 940.595696] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224695, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071011} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.596633] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.598702] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866e5cb2-ba0f-48f6-9302-5cf7f81fcf3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.627300] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] e6435156-fb2e-4912-9587-99eeafeded87/e6435156-fb2e-4912-9587-99eeafeded87.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.628456] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abafaa9e-3648-43e4-b4f1-f767adc156f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.649759] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 940.649759] env[61947]: value = "task-1224696" [ 940.649759] env[61947]: _type = "Task" [ 940.649759] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.659165] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224696, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.703353] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Successfully updated port: fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.010389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.010786] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.011103] env[61947]: DEBUG nova.compute.manager [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Going to confirm migration 3 {{(pid=61947) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 941.083948] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.084305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.084343] env[61947]: DEBUG nova.network.neutron [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.160484] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224696, 'name': ReconfigVM_Task, 'duration_secs': 0.307509} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.160783] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Reconfigured VM instance instance-0000004f to attach disk [datastore1] e6435156-fb2e-4912-9587-99eeafeded87/e6435156-fb2e-4912-9587-99eeafeded87.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.161449] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4411a1c-1000-4f9d-9efd-c62859b07446 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.168061] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 941.168061] env[61947]: value = "task-1224697" [ 941.168061] env[61947]: _type = "Task" [ 941.168061] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.176737] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224697, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.207620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.207620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquired lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.207620] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.585614] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.585821] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.586013] env[61947]: DEBUG nova.network.neutron [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.586211] env[61947]: DEBUG nova.objects.instance [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'info_cache' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.682530] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224697, 'name': Rename_Task, 'duration_secs': 0.142313} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.682807] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 941.683070] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bed1df1-7a16-4e30-b9d5-dff4431a3aee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.690404] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 941.690404] env[61947]: value = "task-1224698" [ 941.690404] env[61947]: _type = "Task" [ 941.690404] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.698249] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224698, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.765689] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.789740] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f42456-5d2d-452f-9bfb-f793cc49b405 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.799067] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663dee51-7eef-465c-8c1d-5114a9e53102 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.832672] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3b4feb-d296-41bb-bdbb-15a3afcc2cb5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.840713] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e641d786-7bfb-40e8-82a0-47e5bfdf698c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.856027] env[61947]: DEBUG nova.compute.provider_tree [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.895149] env[61947]: DEBUG nova.network.neutron [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.953503] env[61947]: DEBUG nova.network.neutron [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Updating instance_info_cache with network_info: [{"id": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "address": "fa:16:3e:97:65:4d", "network": {"id": "0d15d07d-4c44-4193-a0da-e237e8d4ea0e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-761890794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2f48256eb99494d942f3bb8c4a62a67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe941a11-be", "ovs_interfaceid": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.200412] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224698, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.359132] env[61947]: DEBUG nova.scheduler.client.report [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.397477] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.456636] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Releasing lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.456943] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Instance network_info: |[{"id": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "address": "fa:16:3e:97:65:4d", "network": {"id": "0d15d07d-4c44-4193-a0da-e237e8d4ea0e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-761890794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2f48256eb99494d942f3bb8c4a62a67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe941a11-be", "ovs_interfaceid": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.457425] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:65:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe941a11-be7a-47ce-b314-cbeab2ec8cd1', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.464930] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Creating folder: Project (b2f48256eb99494d942f3bb8c4a62a67). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 942.465239] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9f1fae7-e460-4f1f-90df-5a7d9eb6e893 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.476065] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Created folder: Project (b2f48256eb99494d942f3bb8c4a62a67) in parent group-v264556. [ 942.476267] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Creating folder: Instances. Parent ref: group-v264697. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 942.476512] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e35fd23-cf94-4cae-99b8-857542310150 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.484965] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Created folder: Instances in parent group-v264697. [ 942.485221] env[61947]: DEBUG oslo.service.loopingcall [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.485410] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.485612] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ffca9c0-fefd-4fc5-b4d1-ee78ba322d62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.506256] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.506256] env[61947]: value = "task-1224701" [ 942.506256] env[61947]: _type = "Task" [ 942.506256] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.513840] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224701, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.598176] env[61947]: DEBUG nova.compute.manager [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Received event network-changed-fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 942.598396] env[61947]: DEBUG nova.compute.manager [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Refreshing instance network info cache due to event network-changed-fe941a11-be7a-47ce-b314-cbeab2ec8cd1. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 942.598618] env[61947]: DEBUG oslo_concurrency.lockutils [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] Acquiring lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.598851] env[61947]: DEBUG oslo_concurrency.lockutils [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] Acquired lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.599096] env[61947]: DEBUG nova.network.neutron [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Refreshing network info cache for port fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.701551] env[61947]: DEBUG oslo_vmware.api [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224698, 'name': PowerOnVM_Task, 'duration_secs': 0.627171} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.701887] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 942.702128] env[61947]: INFO nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Took 8.43 seconds to spawn the instance on the hypervisor. [ 942.702322] env[61947]: DEBUG nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.703149] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3600021-c06f-4df4-9fd3-debc9059feea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.847329] env[61947]: DEBUG nova.network.neutron [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.864090] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.864630] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.867143] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.191s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.869218] env[61947]: INFO nova.compute.claims [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.015979] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224701, 'name': CreateVM_Task, 'duration_secs': 0.393974} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.016203] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.016907] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.017093] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.017431] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.017706] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-217b24be-0dfe-4fac-ac25-5fbd16774d4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.021809] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 943.021809] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5230b5c4-4979-985c-dfcd-75895d33a5c3" [ 943.021809] env[61947]: _type = "Task" [ 943.021809] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.029304] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5230b5c4-4979-985c-dfcd-75895d33a5c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.223045] env[61947]: INFO nova.compute.manager [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Took 14.32 seconds to build instance. [ 943.382021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.404253] env[61947]: DEBUG nova.objects.instance [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'migration_context' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.409445] env[61947]: DEBUG nova.compute.utils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.508499] env[61947]: DEBUG nova.objects.base [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.508863] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.509027] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 943.518027] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904adc31-157a-45c9-9331-0350c9d36750 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.558970] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5230b5c4-4979-985c-dfcd-75895d33a5c3, 'name': SearchDatastore_Task, 'duration_secs': 0.007415} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.560271] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.560557] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.560744] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.560897] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.561134] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.561356] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccd0af2f-a133-4941-b0f4-95d7b013ae9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.564073] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cea08a28-b388-4875-b275-72a603b71a90 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.569854] env[61947]: DEBUG oslo_vmware.api [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 943.569854] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a77fd0-d6e6-41f2-82e0-b2291513d683" [ 943.569854] env[61947]: _type = "Task" [ 943.569854] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.574334] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.574518] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.575898] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39a28740-3eff-4543-b365-6eebbe516fc6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.581709] env[61947]: DEBUG oslo_vmware.api [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a77fd0-d6e6-41f2-82e0-b2291513d683, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.584487] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 943.584487] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5286ee5d-cd32-7f15-b601-069973f371b9" [ 943.584487] env[61947]: _type = "Task" [ 943.584487] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.592163] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5286ee5d-cd32-7f15-b601-069973f371b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.622159] env[61947]: DEBUG nova.policy [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 943.688170] env[61947]: DEBUG nova.network.neutron [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Updated VIF entry in instance network info cache for port fe941a11-be7a-47ce-b314-cbeab2ec8cd1. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.688289] env[61947]: DEBUG nova.network.neutron [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Updating instance_info_cache with network_info: [{"id": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "address": "fa:16:3e:97:65:4d", "network": {"id": "0d15d07d-4c44-4193-a0da-e237e8d4ea0e", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-761890794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2f48256eb99494d942f3bb8c4a62a67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe941a11-be", "ovs_interfaceid": "fe941a11-be7a-47ce-b314-cbeab2ec8cd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.725413] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8419b03-1a6f-4696-a6dc-9431a545df37 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.854s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.951695] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Successfully created port: e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.001813] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.075098] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06016b50-f569-4d0a-a6e3-d5e2e4f94a06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.098939] env[61947]: DEBUG oslo_vmware.api [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a77fd0-d6e6-41f2-82e0-b2291513d683, 'name': SearchDatastore_Task, 'duration_secs': 0.0144} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.099298] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.109344] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.115461] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5286ee5d-cd32-7f15-b601-069973f371b9, 'name': SearchDatastore_Task, 'duration_secs': 0.007714} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.116346] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ef30f3-a2d6-492b-8c37-dea47a42acd6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.122963] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 944.122963] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52eff059-22af-f2ad-87f6-348cc914b8f6" [ 944.122963] env[61947]: _type = "Task" [ 944.122963] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.131280] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52eff059-22af-f2ad-87f6-348cc914b8f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.191565] env[61947]: DEBUG oslo_concurrency.lockutils [req-cde0f72a-1795-466c-a231-cf3da4a4a729 req-c42d0272-655b-4224-a893-602f6dafe111 service nova] Releasing lock "refresh_cache-16dd2b82-1925-410b-b3b8-d3e2cf8648b0" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.245786] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab001795-f02b-458c-a575-e5ae954cae7e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.253710] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dbf45d-49d8-48aa-9a4f-7cc91aed121b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.285335] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0c250d-b55e-4b9d-962c-ec4b9b27d7d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.292929] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cebf26-fcf8-44e0-b6c1-a4cb9ada9cd4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.308163] env[61947]: DEBUG nova.compute.provider_tree [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.604972] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 944.605306] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-911d7453-62bd-47ab-8d06-48e0c6dd4dbf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.612863] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 944.612863] env[61947]: value = "task-1224702" [ 944.612863] env[61947]: _type = "Task" [ 944.612863] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.621079] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224702, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.632695] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52eff059-22af-f2ad-87f6-348cc914b8f6, 'name': SearchDatastore_Task, 'duration_secs': 0.008747} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.632887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.633156] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 16dd2b82-1925-410b-b3b8-d3e2cf8648b0/16dd2b82-1925-410b-b3b8-d3e2cf8648b0.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.633440] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f8dbb47-dfa5-41cf-b934-a6bfe5375347 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.639166] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 944.639166] env[61947]: value = "task-1224703" [ 944.639166] env[61947]: _type = "Task" [ 944.639166] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.647126] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.811559] env[61947]: DEBUG nova.scheduler.client.report [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.016129] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.043643] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.043965] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.044205] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.044438] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.044629] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.044898] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.045213] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.045464] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.045733] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.046009] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.046259] env[61947]: DEBUG nova.virt.hardware [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.047291] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5d5ef7-b8c0-4a44-8aac-d9d90f19d7f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.059591] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd2987f-23d7-432d-a6f7-e15469547069 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.125465] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224702, 'name': PowerOffVM_Task, 'duration_secs': 0.187616} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.125835] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 945.126215] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 945.150306] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503673} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.150497] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 16dd2b82-1925-410b-b3b8-d3e2cf8648b0/16dd2b82-1925-410b-b3b8-d3e2cf8648b0.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.150737] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.151042] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb9e3633-8897-4223-9be2-d8bed4a52cb5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.159439] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 945.159439] env[61947]: value = "task-1224704" [ 945.159439] env[61947]: _type = "Task" [ 945.159439] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.169768] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.229838] env[61947]: DEBUG nova.compute.manager [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.230322] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45596160-f4aa-470e-bad6-331dd5c68d30 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.316424] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.317008] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.319487] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.702s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.319669] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.320541] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 945.320541] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.293s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.321982] env[61947]: INFO nova.compute.claims [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.325132] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c345ed2-ac4d-4497-b85d-dd62cdcd0c2e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.333493] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c882d5eb-d7bd-410f-8a52-28c4edd615e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.347626] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409c44bd-c456-4d52-b8e1-71cbc85f43ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.354995] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee1928f-c077-458d-80c7-0295ef18e7bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.385364] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179928MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 945.385663] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.419624] env[61947]: DEBUG nova.compute.manager [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Received event network-vif-plugged-e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.419862] env[61947]: DEBUG oslo_concurrency.lockutils [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] Acquiring lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.420076] env[61947]: DEBUG oslo_concurrency.lockutils [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] Lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.420306] env[61947]: DEBUG oslo_concurrency.lockutils [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] Lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.420414] env[61947]: DEBUG nova.compute.manager [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] No waiting events found dispatching network-vif-plugged-e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.420575] env[61947]: WARNING nova.compute.manager [req-b26d9c91-ff29-4b4b-8acb-891bf289efd5 req-a100be27-ba8b-4b3e-900a-f34549c3db61 service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Received unexpected event network-vif-plugged-e75bf013-ccbc-4dc2-ad7f-efa662784f0c for instance with vm_state building and task_state spawning. [ 945.505033] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Successfully updated port: e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.634804] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.635225] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.635502] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.635824] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.636111] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.636320] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.636591] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.636784] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.636967] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.637153] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.637395] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.643031] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adddfc8f-8064-4230-9203-3f92b9b04118 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.661193] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 945.661193] env[61947]: value = "task-1224705" [ 945.661193] env[61947]: _type = "Task" [ 945.661193] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.672463] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098314} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.675736] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.676100] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224705, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.676819] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5da80d-2473-49e2-8357-1f0f3c9c3e04 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.701196] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 16dd2b82-1925-410b-b3b8-d3e2cf8648b0/16dd2b82-1925-410b-b3b8-d3e2cf8648b0.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.701600] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f9ab072-7ae8-4db9-9895-b91525b04685 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.721392] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 945.721392] env[61947]: value = "task-1224706" [ 945.721392] env[61947]: _type = "Task" [ 945.721392] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.729480] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.740796] env[61947]: INFO nova.compute.manager [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] instance snapshotting [ 945.743732] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a302c5c4-ef95-46da-90c4-ad7c2992f119 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.761641] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2fc34a-9051-4a63-a1b5-7393b3851deb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.827250] env[61947]: DEBUG nova.compute.utils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.829561] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.829699] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.906771] env[61947]: DEBUG nova.policy [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd184f9216c164f54ae07005cb607109d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9ffe92d54eb4e7d91e95f68f195c49e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 946.008264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.008264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.008264] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 946.175773] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224705, 'name': ReconfigVM_Task, 'duration_secs': 0.510409} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.176129] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 946.231447] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.272564] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 946.273037] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-25ef057d-f0cd-4c4b-a26b-6fcb0ef0bd5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.276805] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Successfully created port: 013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.284464] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 946.284464] env[61947]: value = "task-1224707" [ 946.284464] env[61947]: _type = "Task" [ 946.284464] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.292817] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224707, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.337201] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.551918] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0e72e4-0660-486f-88a0-2cc2795ada34 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.555215] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.561893] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a28346-c395-4bd8-93a3-25a05d5dcc46 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.593798] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d586568-f970-4951-a673-499130c65535 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.601080] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bb99be-3deb-4342-bc42-b4e6fec60669 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.614347] env[61947]: DEBUG nova.compute.provider_tree [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.682465] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.682680] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.683034] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.683313] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.683951] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.683951] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.683951] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.684127] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.684294] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.684490] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.684846] env[61947]: DEBUG nova.virt.hardware [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.690737] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 946.691436] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b64636b-6f0f-4d6b-8f85-7514ec67d91c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.712485] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 946.712485] env[61947]: value = "task-1224708" [ 946.712485] env[61947]: _type = "Task" [ 946.712485] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.723415] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224708, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.731335] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224706, 'name': ReconfigVM_Task, 'duration_secs': 0.695909} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.731596] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 16dd2b82-1925-410b-b3b8-d3e2cf8648b0/16dd2b82-1925-410b-b3b8-d3e2cf8648b0.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.732792] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c24e5e1e-cabc-4f4a-8f6b-931e96fd7359 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.738458] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 946.738458] env[61947]: value = "task-1224709" [ 946.738458] env[61947]: _type = "Task" [ 946.738458] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.746287] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224709, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.780807] env[61947]: DEBUG nova.network.neutron [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Updating instance_info_cache with network_info: [{"id": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "address": "fa:16:3e:56:61:18", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape75bf013-cc", "ovs_interfaceid": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.794045] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224707, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.117683] env[61947]: DEBUG nova.scheduler.client.report [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.223339] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224708, 'name': ReconfigVM_Task, 'duration_secs': 0.232082} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.224044] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 947.224438] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e260536d-c640-49c2-898c-5ef3c120aca5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.247212] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.250355] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7de344d0-2a14-4c89-82b1-9c30cb326960 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.268159] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224709, 'name': Rename_Task, 'duration_secs': 0.213612} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.269390] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.269701] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 947.269701] env[61947]: value = "task-1224710" [ 947.269701] env[61947]: _type = "Task" [ 947.269701] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.269892] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40e79dbd-41bf-48c1-8e1c-bcf545af3706 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.279168] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.280303] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 947.280303] env[61947]: value = "task-1224711" [ 947.280303] env[61947]: _type = "Task" [ 947.280303] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.283228] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.283510] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Instance network_info: |[{"id": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "address": "fa:16:3e:56:61:18", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape75bf013-cc", "ovs_interfaceid": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 947.283873] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:61:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e75bf013-ccbc-4dc2-ad7f-efa662784f0c', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.291381] env[61947]: DEBUG oslo.service.loopingcall [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.291980] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 947.295112] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c88ef8e3-17f3-4c6a-af9a-2121a03ff8da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.312971] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224711, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.319325] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224707, 'name': CreateSnapshot_Task, 'duration_secs': 0.577097} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.320435] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 947.320778] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.320778] env[61947]: value = "task-1224712" [ 947.320778] env[61947]: _type = "Task" [ 947.320778] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.321463] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddd3cfd-2ef1-440d-b806-1ef7205fd387 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.338628] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224712, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.349905] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.376871] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.377196] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.377329] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.377514] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.377668] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.377857] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.378095] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.378268] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.378486] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.378738] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.378938] env[61947]: DEBUG nova.virt.hardware [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.380133] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2760bc19-22a5-4a59-9cb5-3a9bc702ec2f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.388155] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808cadd1-8053-4baf-833e-dff46ed15dee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.446995] env[61947]: DEBUG nova.compute.manager [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Received event network-changed-e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.447220] env[61947]: DEBUG nova.compute.manager [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Refreshing instance network info cache due to event network-changed-e75bf013-ccbc-4dc2-ad7f-efa662784f0c. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 947.447441] env[61947]: DEBUG oslo_concurrency.lockutils [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] Acquiring lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.447590] env[61947]: DEBUG oslo_concurrency.lockutils [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] Acquired lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.447755] env[61947]: DEBUG nova.network.neutron [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Refreshing network info cache for port e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.624054] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.624205] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.627296] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.518s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.781585] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224710, 'name': ReconfigVM_Task, 'duration_secs': 0.34803} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.788756] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Reconfigured VM instance instance-0000004d to attach disk [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2/dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.789126] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 947.793523] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Successfully updated port: 013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.800487] env[61947]: DEBUG oslo_vmware.api [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224711, 'name': PowerOnVM_Task, 'duration_secs': 0.490658} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.800784] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.801044] env[61947]: INFO nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Took 7.54 seconds to spawn the instance on the hypervisor. [ 947.801273] env[61947]: DEBUG nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.802254] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b541c3-a569-4cc3-89d6-123dbbd16f6b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.833396] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224712, 'name': CreateVM_Task, 'duration_secs': 0.320692} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.833565] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 947.834307] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.834478] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.834803] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.835066] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5517560e-f6b2-4805-8bfb-8ea14febfb6b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.843023] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 947.843912] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2f46d038-d340-4f70-810e-019e448cbe62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.849954] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 947.849954] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523304be-b094-3327-b975-f529742c4a08" [ 947.849954] env[61947]: _type = "Task" [ 947.849954] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.854449] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 947.854449] env[61947]: value = "task-1224713" [ 947.854449] env[61947]: _type = "Task" [ 947.854449] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.860849] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523304be-b094-3327-b975-f529742c4a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.865771] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224713, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.132959] env[61947]: DEBUG nova.compute.utils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.136137] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.139861] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.192219] env[61947]: DEBUG nova.network.neutron [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Updated VIF entry in instance network info cache for port e75bf013-ccbc-4dc2-ad7f-efa662784f0c. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.192715] env[61947]: DEBUG nova.network.neutron [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Updating instance_info_cache with network_info: [{"id": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "address": "fa:16:3e:56:61:18", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape75bf013-cc", "ovs_interfaceid": "e75bf013-ccbc-4dc2-ad7f-efa662784f0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.229330] env[61947]: DEBUG nova.policy [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 948.298432] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.298602] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.298841] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.300762] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8ba622-1315-4d5a-a321-81794c05f4b0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.335963] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae057fdf-b369-454a-a457-9fce168b18e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.341263] env[61947]: INFO nova.compute.manager [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Took 18.02 seconds to build instance. [ 948.357813] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 948.377730] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523304be-b094-3327-b975-f529742c4a08, 'name': SearchDatastore_Task, 'duration_secs': 0.029256} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.381301] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.382168] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.382168] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.382168] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.382168] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.382615] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224713, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.385337] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f6c8677-ba72-4658-9699-f114ded61d22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.394742] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.394988] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 948.395907] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57ae9984-9a85-4f2f-915b-54aca2aba33b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.404680] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 948.404680] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520b9905-a13a-9363-498f-12b2a6be796f" [ 948.404680] env[61947]: _type = "Task" [ 948.404680] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.414258] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520b9905-a13a-9363-498f-12b2a6be796f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.430868] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c677c5dd-0f0f-46be-a80a-822ea1ef5452 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.439468] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8185d0be-f3b6-4d2f-b5bc-4b96fe487292 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.473201] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33a13d3-10c2-4600-9195-619de8b9e682 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.481457] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b757419f-c8dc-4d07-9d0c-18532a19cd57 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.495886] env[61947]: DEBUG nova.compute.provider_tree [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.637370] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.697560] env[61947]: DEBUG oslo_concurrency.lockutils [req-38779644-5bb5-4271-ba76-3e99a890e46b req-b4abcc9a-4d57-474d-aa12-6c34a6c9676b service nova] Releasing lock "refresh_cache-ab21547e-0ddb-4271-967f-bea5056c163a" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.705145] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Successfully created port: c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.843459] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0238ba6b-3b11-4f03-9d69-92028a57eb3e tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.533s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.846771] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.867912] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224713, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.917447] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520b9905-a13a-9363-498f-12b2a6be796f, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.918384] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cef1a591-5c49-42ed-a74c-dcdeed4512b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.927674] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 948.927674] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7acc2-ed95-80fb-235e-85b0b512c115" [ 948.927674] env[61947]: _type = "Task" [ 948.927674] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.937401] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7acc2-ed95-80fb-235e-85b0b512c115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.971714] env[61947]: DEBUG nova.network.neutron [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Port 3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 949.000529] env[61947]: DEBUG nova.scheduler.client.report [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.120792] env[61947]: DEBUG nova.network.neutron [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Updating instance_info_cache with network_info: [{"id": "013a9067-5135-456a-812a-791fd1e20a3b", "address": "fa:16:3e:d6:ef:95", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013a9067-51", "ovs_interfaceid": "013a9067-5135-456a-812a-791fd1e20a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.374336] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224713, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.437406] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7acc2-ed95-80fb-235e-85b0b512c115, 'name': SearchDatastore_Task, 'duration_secs': 0.011405} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.437693] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.437962] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] ab21547e-0ddb-4271-967f-bea5056c163a/ab21547e-0ddb-4271-967f-bea5056c163a.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 949.438238] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-966a8a48-3a41-4adc-bc79-c8fe9590c690 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.445445] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 949.445445] env[61947]: value = "task-1224714" [ 949.445445] env[61947]: _type = "Task" [ 949.445445] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.455195] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.472855] env[61947]: DEBUG nova.compute.manager [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Received event network-vif-plugged-013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.473219] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Acquiring lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.473542] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.473840] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.474162] env[61947]: DEBUG nova.compute.manager [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] No waiting events found dispatching network-vif-plugged-013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.474404] env[61947]: WARNING nova.compute.manager [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Received unexpected event network-vif-plugged-013a9067-5135-456a-812a-791fd1e20a3b for instance with vm_state building and task_state spawning. [ 949.474583] env[61947]: DEBUG nova.compute.manager [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Received event network-changed-013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.474746] env[61947]: DEBUG nova.compute.manager [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Refreshing instance network info cache due to event network-changed-013a9067-5135-456a-812a-791fd1e20a3b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 949.474927] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Acquiring lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.521585] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.521886] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.522150] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.522359] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.522554] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.524860] env[61947]: INFO nova.compute.manager [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Terminating instance [ 949.527235] env[61947]: DEBUG nova.compute.manager [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.527393] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.528370] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187c2d75-3921-4363-9cca-14aaf9e8780c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.537511] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.537778] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c7971cc-875d-4c0b-b5a8-be9feae02786 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.545332] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 949.545332] env[61947]: value = "task-1224715" [ 949.545332] env[61947]: _type = "Task" [ 949.545332] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.555468] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.627704] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.628126] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Instance network_info: |[{"id": "013a9067-5135-456a-812a-791fd1e20a3b", "address": "fa:16:3e:d6:ef:95", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013a9067-51", "ovs_interfaceid": "013a9067-5135-456a-812a-791fd1e20a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.628435] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Acquired lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.628642] env[61947]: DEBUG nova.network.neutron [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Refreshing network info cache for port 013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.630125] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:ef:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '517421c3-bea0-419c-ab0b-987815e5d160', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '013a9067-5135-456a-812a-791fd1e20a3b', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.638347] env[61947]: DEBUG oslo.service.loopingcall [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.639580] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 949.639864] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-365fbc77-1ba6-46b1-9ab6-34ab20ada2c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.657678] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.666667] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.666667] env[61947]: value = "task-1224716" [ 949.666667] env[61947]: _type = "Task" [ 949.666667] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.675925] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224716, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.694639] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.694969] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.695221] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.695455] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.695671] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.695862] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.696129] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.696390] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.696624] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.696832] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.697101] env[61947]: DEBUG nova.virt.hardware [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.698606] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae00fcbe-fdef-43d7-ad18-af3d0f40f609 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.707808] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15208dac-6a32-40fa-b9c2-5bc9a1c8c181 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.872650] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224713, 'name': CloneVM_Task, 'duration_secs': 1.709893} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.873017] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Created linked-clone VM from snapshot [ 949.873875] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9487c13b-033e-4492-82cd-51b73ab25481 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.883547] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Uploading image 15da8d7b-1bef-49dc-9ea4-cb2307155b5f {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 949.895164] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 949.895525] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1a5238da-4e45-4b76-9841-8f549e64c73d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.904453] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 949.904453] env[61947]: value = "task-1224717" [ 949.904453] env[61947]: _type = "Task" [ 949.904453] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.913274] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224717, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.958055] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480952} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.958458] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] ab21547e-0ddb-4271-967f-bea5056c163a/ab21547e-0ddb-4271-967f-bea5056c163a.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 949.958826] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.959161] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0714a850-3076-45d8-8692-d9f1be9e8050 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.966015] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 949.966015] env[61947]: value = "task-1224718" [ 949.966015] env[61947]: _type = "Task" [ 949.966015] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.974429] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.998095] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.998355] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.998530] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.010849] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.383s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.011161] env[61947]: DEBUG nova.compute.manager [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61947) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 950.014445] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.629s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.056293] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224715, 'name': PowerOffVM_Task, 'duration_secs': 0.437764} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.057042] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.057042] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.057232] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76c56528-89db-4558-860c-9875a04b87a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.129656] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.129924] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.130142] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Deleting the datastore file [datastore1] 16dd2b82-1925-410b-b3b8-d3e2cf8648b0 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.130747] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-002c1ddc-7d33-4561-8b07-0ce9d7b0e31b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.138540] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for the task: (returnval){ [ 950.138540] env[61947]: value = "task-1224720" [ 950.138540] env[61947]: _type = "Task" [ 950.138540] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.148235] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224720, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.176721] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224716, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.413997] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224717, 'name': Destroy_Task, 'duration_secs': 0.322003} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.414274] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Destroyed the VM [ 950.414498] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 950.414758] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-83170f5f-c94e-45a8-8f23-3f1ca498111d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.421295] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 950.421295] env[61947]: value = "task-1224721" [ 950.421295] env[61947]: _type = "Task" [ 950.421295] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.428863] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224721, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.475410] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06058} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.475694] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.476512] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad61e6e0-50e4-47d8-a5bb-7f14eb80299e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.499945] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] ab21547e-0ddb-4271-967f-bea5056c163a/ab21547e-0ddb-4271-967f-bea5056c163a.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.505024] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0680386-2604-4d16-8f52-fffa8d0103ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.539251] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 950.539251] env[61947]: value = "task-1224722" [ 950.539251] env[61947]: _type = "Task" [ 950.539251] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.543705] env[61947]: DEBUG nova.network.neutron [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Updated VIF entry in instance network info cache for port 013a9067-5135-456a-812a-791fd1e20a3b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.544415] env[61947]: DEBUG nova.network.neutron [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Updating instance_info_cache with network_info: [{"id": "013a9067-5135-456a-812a-791fd1e20a3b", "address": "fa:16:3e:d6:ef:95", "network": {"id": "96a35541-e288-4940-9af4-6f6cf0994d2d", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1505444194-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9ffe92d54eb4e7d91e95f68f195c49e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013a9067-51", "ovs_interfaceid": "013a9067-5135-456a-812a-791fd1e20a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.552912] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224722, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.595870] env[61947]: INFO nova.scheduler.client.report [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted allocation for migration 33113961-17ce-41e5-b039-4af1e078f3ef [ 950.649934] env[61947]: DEBUG oslo_vmware.api [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Task: {'id': task-1224720, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160103} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.650226] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.650416] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.650595] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.650772] env[61947]: INFO nova.compute.manager [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 950.651034] env[61947]: DEBUG oslo.service.loopingcall [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.651236] env[61947]: DEBUG nova.compute.manager [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.651332] env[61947]: DEBUG nova.network.neutron [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.677591] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224716, 'name': CreateVM_Task, 'duration_secs': 0.515758} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.677761] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 950.678448] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.678620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.678998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.679537] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-742db313-f073-4213-b9c4-264d6ff605b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.684224] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 950.684224] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523ded8a-ae67-600c-d71b-5428ef8e9dc7" [ 950.684224] env[61947]: _type = "Task" [ 950.684224] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.688054] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Successfully updated port: c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.694189] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523ded8a-ae67-600c-d71b-5428ef8e9dc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.931764] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224721, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.032447] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Applying migration context for instance dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 as it has an incoming, in-progress migration 6a038d9b-8359-4af1-a990-eab5cb334626. Migration status is post-migrating {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 951.034082] env[61947]: INFO nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating resource usage from migration 6a038d9b-8359-4af1-a990-eab5cb334626 [ 951.049155] env[61947]: DEBUG oslo_concurrency.lockutils [req-27f26d49-e08e-4e29-9cbc-309a0412e656 req-f0c36a16-d757-4004-96dd-df080a539799 service nova] Releasing lock "refresh_cache-d2d094c4-0237-48cd-b98b-8d6c49ee5d94" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.049594] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224722, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.050527] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.050643] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8c342d64-75f6-4b7d-829d-889eb3b6122f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.050769] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 21bb0270-bc20-4ec1-9599-d676845b0dc7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.050996] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.050996] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 91205e54-6bcb-403c-8308-0f0692fe4fba actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051131] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance c4647324-7e52-44d4-a4b0-961684cc1c4c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051232] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ab027a9d-f51d-4cb6-8aaf-ebd507548657 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051343] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance e6435156-fb2e-4912-9587-99eeafeded87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051454] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 16dd2b82-1925-410b-b3b8-d3e2cf8648b0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051565] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Migration 6a038d9b-8359-4af1-a990-eab5cb334626 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 951.051675] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051783] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ab21547e-0ddb-4271-967f-bea5056c163a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.051892] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d2d094c4-0237-48cd-b98b-8d6c49ee5d94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.052177] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance fcd2e134-4d5f-43d5-a09c-55a49f8ce48b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 951.052334] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 951.052478] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3328MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 951.070810] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.071007] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.071205] env[61947]: DEBUG nova.network.neutron [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.103495] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3f247af-7cd9-47a1-b2b5-97c0a12a472b tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.093s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.192235] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.192731] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.192890] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.203092] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523ded8a-ae67-600c-d71b-5428ef8e9dc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009419} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.203967] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.204679] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.204679] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.204832] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.205086] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.205816] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-614cb49a-109b-4702-911f-4eb21f31a62f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.214753] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.214959] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.218393] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73706d65-5e90-485d-9b6e-56089f743898 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.224667] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 951.224667] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281a6dd-6b0c-17b4-9fd8-d61ab8ec1580" [ 951.224667] env[61947]: _type = "Task" [ 951.224667] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.235023] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281a6dd-6b0c-17b4-9fd8-d61ab8ec1580, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.249125] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff54567-3e6d-432a-8756-a736cc025445 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.256571] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea144fe-0b67-4ba1-a568-9573a4cefb17 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.288834] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca58415e-d28c-4df2-a08e-950e7ca9c338 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.296639] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce9f168-16ad-4a80-9932-b297d9afbf19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.309760] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.413056] env[61947]: DEBUG nova.network.neutron [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.434355] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224721, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.441666] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.441958] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.520596] env[61947]: DEBUG nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-vif-plugged-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.520833] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.521062] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.521283] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.521472] env[61947]: DEBUG nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] No waiting events found dispatching network-vif-plugged-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.521660] env[61947]: WARNING nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received unexpected event network-vif-plugged-c32a51a8-436f-4d1d-abb7-452a152efd91 for instance with vm_state building and task_state spawning. [ 951.521860] env[61947]: DEBUG nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.522043] env[61947]: DEBUG nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.522218] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.550113] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224722, 'name': ReconfigVM_Task, 'duration_secs': 0.975528} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.550403] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Reconfigured VM instance instance-00000051 to attach disk [datastore1] ab21547e-0ddb-4271-967f-bea5056c163a/ab21547e-0ddb-4271-967f-bea5056c163a.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.551047] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d24e9eba-5ee5-428a-8b1d-06454cf3a48d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.557186] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 951.557186] env[61947]: value = "task-1224723" [ 951.557186] env[61947]: _type = "Task" [ 951.557186] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.565289] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224723, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.695519] env[61947]: DEBUG nova.objects.instance [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.738998] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281a6dd-6b0c-17b4-9fd8-d61ab8ec1580, 'name': SearchDatastore_Task, 'duration_secs': 0.008658} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.739818] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2365bfc9-d324-4f04-adc2-d9f6ca338559 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.745182] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 951.745182] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522ef050-2a2d-daff-1fa9-a0cb9987d6c6" [ 951.745182] env[61947]: _type = "Task" [ 951.745182] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.745834] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.767277] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522ef050-2a2d-daff-1fa9-a0cb9987d6c6, 'name': SearchDatastore_Task, 'duration_secs': 0.019049} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.767538] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.767799] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] d2d094c4-0237-48cd-b98b-8d6c49ee5d94/d2d094c4-0237-48cd-b98b-8d6c49ee5d94.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 951.768079] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d48c441-3f5b-4167-9b1a-bc2520ecc05f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.775381] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 951.775381] env[61947]: value = "task-1224724" [ 951.775381] env[61947]: _type = "Task" [ 951.775381] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.784375] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.813363] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.862389] env[61947]: DEBUG nova.network.neutron [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.915568] env[61947]: INFO nova.compute.manager [-] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Took 1.26 seconds to deallocate network for instance. [ 951.928485] env[61947]: DEBUG nova.network.neutron [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.935630] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224721, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.944906] env[61947]: DEBUG nova.compute.utils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.068987] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224723, 'name': Rename_Task, 'duration_secs': 0.131969} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.069439] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 952.069821] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-721f8a64-0e59-41d8-ba5f-6883ee3a43b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.078847] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 952.078847] env[61947]: value = "task-1224725" [ 952.078847] env[61947]: _type = "Task" [ 952.078847] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.090063] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.201721] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.202061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.202345] env[61947]: DEBUG nova.network.neutron [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.202569] env[61947]: DEBUG nova.objects.instance [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'info_cache' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.286074] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484104} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.286363] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] d2d094c4-0237-48cd-b98b-8d6c49ee5d94/d2d094c4-0237-48cd-b98b-8d6c49ee5d94.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 952.286592] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.286864] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83528dec-d9c5-49da-933b-9dd94ef2591b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.292782] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 952.292782] env[61947]: value = "task-1224726" [ 952.292782] env[61947]: _type = "Task" [ 952.292782] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.300285] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.318128] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 952.318329] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.304s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.364773] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.423321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.423616] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.423852] env[61947]: DEBUG nova.objects.instance [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lazy-loading 'resources' on Instance uuid 16dd2b82-1925-410b-b3b8-d3e2cf8648b0 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.431328] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.431641] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Instance network_info: |[{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.434849] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.435103] env[61947]: DEBUG nova.network.neutron [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.436185] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:87:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c32a51a8-436f-4d1d-abb7-452a152efd91', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.443580] env[61947]: DEBUG oslo.service.loopingcall [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.444658] env[61947]: DEBUG oslo_vmware.api [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224721, 'name': RemoveSnapshot_Task, 'duration_secs': 1.558865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.447046] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 952.447144] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 952.451307] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14461251-577f-4f6d-b6a7-8229fe234131 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.464363] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.022s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.470418] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.470418] env[61947]: value = "task-1224727" [ 952.470418] env[61947]: _type = "Task" [ 952.470418] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.478190] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224727, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.590993] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224725, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.666146] env[61947]: DEBUG nova.network.neutron [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.666592] env[61947]: DEBUG nova.network.neutron [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.706159] env[61947]: DEBUG nova.objects.base [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 952.802878] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150718} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.803240] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.804056] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e099530-4bd0-4952-b217-db8c27affdfc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.830025] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] d2d094c4-0237-48cd-b98b-8d6c49ee5d94/d2d094c4-0237-48cd-b98b-8d6c49ee5d94.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.830025] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeff26a7-8582-489a-96b5-658d674213e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.850731] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 952.850731] env[61947]: value = "task-1224728" [ 952.850731] env[61947]: _type = "Task" [ 952.850731] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.859471] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.890125] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4deb21-d97d-4577-b4f5-e950d170428a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.909315] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc73541-f3e9-4546-9d89-5a36e92c14ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.916011] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 952.968062] env[61947]: WARNING nova.compute.manager [None req-7787d3ab-20e3-4be1-b4e1-ee27bb2246dc tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Image not found during snapshot: nova.exception.ImageNotFound: Image 15da8d7b-1bef-49dc-9ea4-cb2307155b5f could not be found. [ 952.983350] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224727, 'name': CreateVM_Task, 'duration_secs': 0.403241} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.985951] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.987097] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.987185] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.987508] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.987775] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4cd572b-ac33-4a3d-84ba-e943deb2d32a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.992840] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 952.992840] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524835de-87ed-2385-72a9-bacd2619df24" [ 952.992840] env[61947]: _type = "Task" [ 952.992840] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.001940] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524835de-87ed-2385-72a9-bacd2619df24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.093766] env[61947]: DEBUG oslo_vmware.api [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224725, 'name': PowerOnVM_Task, 'duration_secs': 0.626504} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.094156] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 953.094390] env[61947]: INFO nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Took 8.08 seconds to spawn the instance on the hypervisor. [ 953.094595] env[61947]: DEBUG nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.095399] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a245558b-f479-4fd5-a773-872a902cb5c2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.124825] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2214fedb-8a44-4b2a-8276-60670100be24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.132283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1336b9c-08a6-4210-8ea4-de422d2f0081 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.164918] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40344a9-eb99-40f2-9d62-2448e4abc9b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.171938] env[61947]: DEBUG oslo_concurrency.lockutils [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.172212] env[61947]: DEBUG nova.compute.manager [req-5848bcb7-fc53-45f3-9933-cc7842cb75ba req-149a7b04-8edf-422c-bebd-52cf4bf699d8 service nova] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Received event network-vif-deleted-fe941a11-be7a-47ce-b314-cbeab2ec8cd1 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.173619] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11891c5a-e6d6-473b-adae-a34b7001058c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.186778] env[61947]: DEBUG nova.compute.provider_tree [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.361455] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224728, 'name': ReconfigVM_Task, 'duration_secs': 0.304682} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.361731] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Reconfigured VM instance instance-00000053 to attach disk [datastore1] d2d094c4-0237-48cd-b98b-8d6c49ee5d94/d2d094c4-0237-48cd-b98b-8d6c49ee5d94.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.362399] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5774450e-4121-4091-9fff-0c0db9ab11e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.368534] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 953.368534] env[61947]: value = "task-1224729" [ 953.368534] env[61947]: _type = "Task" [ 953.368534] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.375650] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224729, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.416586] env[61947]: DEBUG nova.network.neutron [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [{"id": "46083345-81bb-4fe1-ac51-a4da371e4a16", "address": "fa:16:3e:5e:44:a2", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46083345-81", "ovs_interfaceid": "46083345-81bb-4fe1-ac51-a4da371e4a16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.424794] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 953.425275] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ce2d028-aced-4496-987a-0c72a33e653f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.432683] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 953.432683] env[61947]: value = "task-1224730" [ 953.432683] env[61947]: _type = "Task" [ 953.432683] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.444238] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.503669] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524835de-87ed-2385-72a9-bacd2619df24, 'name': SearchDatastore_Task, 'duration_secs': 0.008637} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.503822] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.504084] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.505081] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.505081] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.505081] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.505081] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be999bae-d69d-41cc-868a-24aa4781f18a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.512453] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.512669] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 953.513522] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da7a4595-1f5a-4e29-b43a-cb3aaf955309 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.518735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.519008] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.519312] env[61947]: INFO nova.compute.manager [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Attaching volume e87cfa37-9ff6-496d-8dc7-6d5aadba8c93 to /dev/sdb [ 953.520931] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 953.520931] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525028d4-b061-22c5-079b-6e849ee96ffb" [ 953.520931] env[61947]: _type = "Task" [ 953.520931] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.530671] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525028d4-b061-22c5-079b-6e849ee96ffb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.553875] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af10beb-c368-4563-b140-05fc4d94cfdc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.561054] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b659d6-dd88-4866-ad8e-ed9dca8f5b40 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.573642] env[61947]: DEBUG nova.virt.block_device [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating existing volume attachment record: fc4b1710-d0eb-4006-b15d-f5df54c90d17 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 953.614823] env[61947]: INFO nova.compute.manager [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Took 20.34 seconds to build instance. [ 953.689474] env[61947]: DEBUG nova.scheduler.client.report [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.877872] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224729, 'name': Rename_Task, 'duration_secs': 0.171251} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.878173] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 953.878431] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aee97dd9-24cc-44db-ae54-cf0afac55365 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.884339] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 953.884339] env[61947]: value = "task-1224734" [ 953.884339] env[61947]: _type = "Task" [ 953.884339] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.891511] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.919727] env[61947]: DEBUG oslo_concurrency.lockutils [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-ab027a9d-f51d-4cb6-8aaf-ebd507548657" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.942863] env[61947]: DEBUG oslo_vmware.api [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224730, 'name': PowerOnVM_Task, 'duration_secs': 0.415678} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.943157] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 953.943348] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e74ff676-549c-42f7-919b-da0deb6779e2 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance 'dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 954.031949] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525028d4-b061-22c5-079b-6e849ee96ffb, 'name': SearchDatastore_Task, 'duration_secs': 0.018198} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.032748] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d04cfff3-9821-49b3-8552-2017e2f25447 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.037624] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 954.037624] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5260f10b-c413-9299-424b-77f6183dcfc7" [ 954.037624] env[61947]: _type = "Task" [ 954.037624] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.045145] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5260f10b-c413-9299-424b-77f6183dcfc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.116487] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f7e0322-e40a-4bb6-878e-88c04cb3177f tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.851s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.134119] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "e6435156-fb2e-4912-9587-99eeafeded87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.134470] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.134722] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "e6435156-fb2e-4912-9587-99eeafeded87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.135531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.135531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.138035] env[61947]: INFO nova.compute.manager [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Terminating instance [ 954.140378] env[61947]: DEBUG nova.compute.manager [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.140643] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 954.141694] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccc5035-eb0c-4d06-a063-563e1544e963 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.152512] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.153070] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-613923bc-6a1d-4843-8423-72075fdafe7c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.159723] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 954.159723] env[61947]: value = "task-1224735" [ 954.159723] env[61947]: _type = "Task" [ 954.159723] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.169031] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.194342] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.219807] env[61947]: INFO nova.scheduler.client.report [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Deleted allocations for instance 16dd2b82-1925-410b-b3b8-d3e2cf8648b0 [ 954.393532] env[61947]: DEBUG oslo_vmware.api [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224734, 'name': PowerOnVM_Task, 'duration_secs': 0.483094} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.393802] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 954.394024] env[61947]: INFO nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Took 7.04 seconds to spawn the instance on the hypervisor. [ 954.394218] env[61947]: DEBUG nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.394981] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32bdd0f-b04d-4f81-a323-f2552fd052d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.423044] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 954.423448] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0918d607-ad0f-4984-8e08-947b56574165 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.431053] env[61947]: DEBUG oslo_vmware.api [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 954.431053] env[61947]: value = "task-1224736" [ 954.431053] env[61947]: _type = "Task" [ 954.431053] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.440205] env[61947]: DEBUG oslo_vmware.api [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224736, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.548359] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5260f10b-c413-9299-424b-77f6183dcfc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.548689] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.548889] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] fcd2e134-4d5f-43d5-a09c-55a49f8ce48b/fcd2e134-4d5f-43d5-a09c-55a49f8ce48b.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 954.549155] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e073dd9f-7c6e-4e04-8cdc-dd58f9c805cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.555812] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 954.555812] env[61947]: value = "task-1224737" [ 954.555812] env[61947]: _type = "Task" [ 954.555812] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.563479] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.668910] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224735, 'name': PowerOffVM_Task, 'duration_secs': 0.196976} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.669218] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 954.669344] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 954.669602] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2ab5da0-61f9-4962-b536-6a591f9d467a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.707800] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "ab21547e-0ddb-4271-967f-bea5056c163a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.708213] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.708471] env[61947]: DEBUG nova.compute.manager [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.709461] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230398d7-abb9-432e-a88e-322ad305b710 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.716241] env[61947]: DEBUG nova.compute.manager [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 954.717065] env[61947]: DEBUG nova.objects.instance [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'flavor' on Instance uuid ab21547e-0ddb-4271-967f-bea5056c163a {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.724791] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 954.724997] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 954.725255] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleting the datastore file [datastore1] e6435156-fb2e-4912-9587-99eeafeded87 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.725911] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7921461d-5d76-40a8-b6c6-f4d285aa335c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.730988] env[61947]: DEBUG oslo_concurrency.lockutils [None req-751289d8-e473-4cfc-a6ae-866128fd104f tempest-ServerMetadataNegativeTestJSON-333006007 tempest-ServerMetadataNegativeTestJSON-333006007-project-member] Lock "16dd2b82-1925-410b-b3b8-d3e2cf8648b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.209s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.738039] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for the task: (returnval){ [ 954.738039] env[61947]: value = "task-1224739" [ 954.738039] env[61947]: _type = "Task" [ 954.738039] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.748611] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.913193] env[61947]: INFO nova.compute.manager [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Took 20.25 seconds to build instance. [ 954.947298] env[61947]: DEBUG oslo_vmware.api [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224736, 'name': PowerOnVM_Task, 'duration_secs': 0.377809} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.947298] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 954.947298] env[61947]: DEBUG nova.compute.manager [None req-cef5976e-81e2-47f4-bbda-948c7be77c98 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.948045] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aad07a4-6678-400c-9763-28e06415cdc7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.065864] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224737, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.223212] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.223493] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51ad6697-6532-4145-947c-6e810efc61d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.230728] env[61947]: DEBUG oslo_vmware.api [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 955.230728] env[61947]: value = "task-1224740" [ 955.230728] env[61947]: _type = "Task" [ 955.230728] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.240577] env[61947]: DEBUG oslo_vmware.api [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.250139] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.415189] env[61947]: DEBUG oslo_concurrency.lockutils [None req-99e32217-ab59-4a6e-9978-e97db4ada910 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.762s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.567211] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575365} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.567505] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] fcd2e134-4d5f-43d5-a09c-55a49f8ce48b/fcd2e134-4d5f-43d5-a09c-55a49f8ce48b.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 955.568338] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.568338] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f93bad4d-9c6f-4bc0-a533-a351f56f5409 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.577185] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 955.577185] env[61947]: value = "task-1224741" [ 955.577185] env[61947]: _type = "Task" [ 955.577185] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.583914] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224741, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.741016] env[61947]: DEBUG oslo_vmware.api [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224740, 'name': PowerOffVM_Task, 'duration_secs': 0.479555} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.744404] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.744635] env[61947]: DEBUG nova.compute.manager [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.745479] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef9641a-ba0e-45e4-a107-50a7cb419fb7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.753380] env[61947]: DEBUG oslo_vmware.api [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Task: {'id': task-1224739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.52753} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.755377] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.755671] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.755919] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.756159] env[61947]: INFO nova.compute.manager [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Took 1.62 seconds to destroy the instance on the hypervisor. [ 955.756468] env[61947]: DEBUG oslo.service.loopingcall [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.758813] env[61947]: DEBUG nova.compute.manager [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.758906] env[61947]: DEBUG nova.network.neutron [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.087975] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224741, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191339} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.092018] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.092018] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e585a1-888e-4812-80b4-a92a865c29c6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.119144] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] fcd2e134-4d5f-43d5-a09c-55a49f8ce48b/fcd2e134-4d5f-43d5-a09c-55a49f8ce48b.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.119633] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ea60f74-f62d-452c-a731-b4c48917c815 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.146685] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 956.146685] env[61947]: value = "task-1224743" [ 956.146685] env[61947]: _type = "Task" [ 956.146685] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.156617] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.269869] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4cb0c4bc-295b-49e1-b0e7-231b31eaab16 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.334880] env[61947]: DEBUG nova.compute.manager [req-03fd85f7-4529-4480-8bd7-106600badd8d req-7358f491-cdfd-4679-a5f0-07cac0171d3f service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Received event network-vif-deleted-2e329412-3c2f-495d-9ed0-0329e52e1c7f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.335044] env[61947]: INFO nova.compute.manager [req-03fd85f7-4529-4480-8bd7-106600badd8d req-7358f491-cdfd-4679-a5f0-07cac0171d3f service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Neutron deleted interface 2e329412-3c2f-495d-9ed0-0329e52e1c7f; detaching it from the instance and deleting it from the info cache [ 956.335907] env[61947]: DEBUG nova.network.neutron [req-03fd85f7-4529-4480-8bd7-106600badd8d req-7358f491-cdfd-4679-a5f0-07cac0171d3f service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.506985] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.507349] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.507578] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.507806] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.508015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.510249] env[61947]: INFO nova.compute.manager [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Terminating instance [ 956.516686] env[61947]: DEBUG nova.compute.manager [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.517019] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 956.517914] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f7b239-00fc-41fc-89e8-3fae487c576f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.527109] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 956.527109] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f203203-e8c2-42c9-82ab-1ef2bdb77dc4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.533563] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 956.533563] env[61947]: value = "task-1224744" [ 956.533563] env[61947]: _type = "Task" [ 956.533563] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.542154] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.656906] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.712096] env[61947]: DEBUG nova.network.neutron [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.838854] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7573bb9d-86ed-491b-92f4-cf7e8c271a95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.849906] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3177502d-6c92-40cd-87b9-15fd5f4400b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.882828] env[61947]: DEBUG nova.compute.manager [req-03fd85f7-4529-4480-8bd7-106600badd8d req-7358f491-cdfd-4679-a5f0-07cac0171d3f service nova] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Detach interface failed, port_id=2e329412-3c2f-495d-9ed0-0329e52e1c7f, reason: Instance e6435156-fb2e-4912-9587-99eeafeded87 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 956.901015] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.901346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.901572] env[61947]: DEBUG nova.compute.manager [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Going to confirm migration 4 {{(pid=61947) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 957.052567] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224744, 'name': PowerOffVM_Task, 'duration_secs': 0.200666} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.052960] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 957.053587] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 957.053901] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6309771-7973-4323-8927-46b702d8ece7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.113564] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 957.113868] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 957.114089] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleting the datastore file [datastore2] ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.114449] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a69dc47-455e-4baf-9681-59290b8bf3a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.120871] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 957.120871] env[61947]: value = "task-1224746" [ 957.120871] env[61947]: _type = "Task" [ 957.120871] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.129658] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.157168] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224743, 'name': ReconfigVM_Task, 'duration_secs': 0.574218} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.157557] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfigured VM instance instance-00000054 to attach disk [datastore1] fcd2e134-4d5f-43d5-a09c-55a49f8ce48b/fcd2e134-4d5f-43d5-a09c-55a49f8ce48b.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.158357] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab8da904-6fc3-406d-988d-f6fa33e40323 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.164664] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 957.164664] env[61947]: value = "task-1224747" [ 957.164664] env[61947]: _type = "Task" [ 957.164664] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.174243] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224747, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.192702] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "ab21547e-0ddb-4271-967f-bea5056c163a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.193074] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.193348] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.193627] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.193885] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.196751] env[61947]: INFO nova.compute.manager [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Terminating instance [ 957.199456] env[61947]: DEBUG nova.compute.manager [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.199456] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.199897] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b993d14b-9ac9-4c24-9b57-857572d5db91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.207208] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 957.207473] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7251e78a-8a0d-4d01-974d-8cfa94076ab6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.217155] env[61947]: INFO nova.compute.manager [-] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Took 1.46 seconds to deallocate network for instance. [ 957.273897] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 957.274093] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 957.274273] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleting the datastore file [datastore1] ab21547e-0ddb-4271-967f-bea5056c163a {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.274578] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-faa2d239-ebd8-4842-a2af-e0fd540a8099 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.281022] env[61947]: DEBUG oslo_vmware.api [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 957.281022] env[61947]: value = "task-1224749" [ 957.281022] env[61947]: _type = "Task" [ 957.281022] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.289184] env[61947]: DEBUG oslo_vmware.api [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.458897] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.459201] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.459441] env[61947]: DEBUG nova.network.neutron [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.459676] env[61947]: DEBUG nova.objects.instance [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lazy-loading 'info_cache' on Instance uuid dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.632101] env[61947]: DEBUG oslo_vmware.api [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18789} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.632101] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.632101] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 957.632101] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 957.632473] env[61947]: INFO nova.compute.manager [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 1.12 seconds to destroy the instance on the hypervisor. [ 957.632473] env[61947]: DEBUG oslo.service.loopingcall [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.632666] env[61947]: DEBUG nova.compute.manager [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.632755] env[61947]: DEBUG nova.network.neutron [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.674926] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224747, 'name': Rename_Task, 'duration_secs': 0.134129} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.675236] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.676051] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f8cade2-b833-46bd-aeb9-482007adab43 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.681333] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 957.681333] env[61947]: value = "task-1224750" [ 957.681333] env[61947]: _type = "Task" [ 957.681333] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.690651] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224750, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.724548] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.724891] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.725162] env[61947]: DEBUG nova.objects.instance [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lazy-loading 'resources' on Instance uuid e6435156-fb2e-4912-9587-99eeafeded87 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.792699] env[61947]: DEBUG oslo_vmware.api [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.120006] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 958.120371] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264706', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'name': 'volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21bb0270-bc20-4ec1-9599-d676845b0dc7', 'attached_at': '', 'detached_at': '', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'serial': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 958.121336] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bbeaed-5335-4f60-97d4-dea79dfe2e97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.139139] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6195260-f742-443e-ac2a-de52bd2e0e1c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.166314] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93/volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.166700] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23ac6429-2d15-4c73-8544-b3a23c1f6d40 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.187539] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 958.187539] env[61947]: value = "task-1224751" [ 958.187539] env[61947]: _type = "Task" [ 958.187539] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.194664] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224750, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.199854] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224751, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.294557] env[61947]: DEBUG oslo_vmware.api [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.556007} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.294920] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.295165] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 958.295380] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 958.295640] env[61947]: INFO nova.compute.manager [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 958.295936] env[61947]: DEBUG oslo.service.loopingcall [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.296178] env[61947]: DEBUG nova.compute.manager [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.296278] env[61947]: DEBUG nova.network.neutron [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.326941] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.327261] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.327506] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.327712] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.327893] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.334242] env[61947]: INFO nova.compute.manager [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Terminating instance [ 958.337644] env[61947]: DEBUG nova.compute.manager [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.337883] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 958.341016] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c59f0bd-2a2b-4896-803f-1fa8bb27e4bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.348548] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 958.348625] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc00f1f9-efbb-43e2-9a3d-01f557f57a7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.357352] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 958.357352] env[61947]: value = "task-1224752" [ 958.357352] env[61947]: _type = "Task" [ 958.357352] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.366805] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.374176] env[61947]: DEBUG nova.compute.manager [req-9be32115-f341-4ae4-9b1f-72e17a9963fe req-d33f0f85-7b66-43f1-8b83-ced1a86aa93f service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Received event network-vif-deleted-46083345-81bb-4fe1-ac51-a4da371e4a16 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.375939] env[61947]: INFO nova.compute.manager [req-9be32115-f341-4ae4-9b1f-72e17a9963fe req-d33f0f85-7b66-43f1-8b83-ced1a86aa93f service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Neutron deleted interface 46083345-81bb-4fe1-ac51-a4da371e4a16; detaching it from the instance and deleting it from the info cache [ 958.375939] env[61947]: DEBUG nova.network.neutron [req-9be32115-f341-4ae4-9b1f-72e17a9963fe req-d33f0f85-7b66-43f1-8b83-ced1a86aa93f service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.454089] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b043934-cdd6-41d2-a471-4ad2ec2e45a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.462185] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8d96d0-847a-4ce3-bda4-593e07613ce8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.494781] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b4aa65-d2de-4737-b6fa-fd55b673e7e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.503790] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1dc230-81a7-439f-a519-ed648ae26597 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.519899] env[61947]: DEBUG nova.compute.provider_tree [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.591161] env[61947]: DEBUG nova.network.neutron [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.701368] env[61947]: DEBUG oslo_vmware.api [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224750, 'name': PowerOnVM_Task, 'duration_secs': 0.734868} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.705009] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.705009] env[61947]: INFO nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Took 9.05 seconds to spawn the instance on the hypervisor. [ 958.705009] env[61947]: DEBUG nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.705173] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224751, 'name': ReconfigVM_Task, 'duration_secs': 0.367317} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.706023] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c4abd1-1b86-429d-8327-63e51a6936de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.708506] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfigured VM instance instance-0000004b to attach disk [datastore2] volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93/volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.714338] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14c7bb92-8cca-498e-bbff-9a81c07e82fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.736191] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 958.736191] env[61947]: value = "task-1224753" [ 958.736191] env[61947]: _type = "Task" [ 958.736191] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.744814] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224753, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.777037] env[61947]: DEBUG nova.network.neutron [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [{"id": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "address": "fa:16:3e:e6:7a:44", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3adc00d1-4b", "ovs_interfaceid": "3adc00d1-4b5f-4857-a2ae-bdc4201f8c69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.872505] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224752, 'name': PowerOffVM_Task, 'duration_secs': 0.450653} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.872505] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 958.872505] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 958.872505] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6357290b-30ce-4bba-9010-5fb06268a921 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.879750] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e96ddf55-9e87-41e1-964f-e44d60f30f5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.890548] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed36f5cc-22a6-4694-b1ee-457b73d3ebbe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.918399] env[61947]: DEBUG nova.compute.manager [req-9be32115-f341-4ae4-9b1f-72e17a9963fe req-d33f0f85-7b66-43f1-8b83-ced1a86aa93f service nova] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Detach interface failed, port_id=46083345-81bb-4fe1-ac51-a4da371e4a16, reason: Instance ab027a9d-f51d-4cb6-8aaf-ebd507548657 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 958.950600] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 958.951020] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 958.952379] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleting the datastore file [datastore1] d2d094c4-0237-48cd-b98b-8d6c49ee5d94 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.952379] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e261a80-4652-4d6b-be49-fe79f52a18f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.959100] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 958.959100] env[61947]: value = "task-1224755" [ 958.959100] env[61947]: _type = "Task" [ 958.959100] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.971796] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.023590] env[61947]: DEBUG nova.scheduler.client.report [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.093814] env[61947]: INFO nova.compute.manager [-] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Took 1.46 seconds to deallocate network for instance. [ 959.132783] env[61947]: DEBUG nova.network.neutron [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.247838] env[61947]: INFO nova.compute.manager [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Took 21.23 seconds to build instance. [ 959.252719] env[61947]: DEBUG oslo_vmware.api [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224753, 'name': ReconfigVM_Task, 'duration_secs': 0.14267} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.253062] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264706', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'name': 'volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21bb0270-bc20-4ec1-9599-d676845b0dc7', 'attached_at': '', 'detached_at': '', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'serial': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 959.281030] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.281223] env[61947]: DEBUG nova.objects.instance [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lazy-loading 'migration_context' on Instance uuid dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.473029] env[61947]: DEBUG oslo_vmware.api [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242043} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.473029] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.473273] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 959.473453] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 959.473533] env[61947]: INFO nova.compute.manager [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Took 1.14 seconds to destroy the instance on the hypervisor. [ 959.473777] env[61947]: DEBUG oslo.service.loopingcall [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.473972] env[61947]: DEBUG nova.compute.manager [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.474083] env[61947]: DEBUG nova.network.neutron [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.530392] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.554288] env[61947]: INFO nova.scheduler.client.report [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Deleted allocations for instance e6435156-fb2e-4912-9587-99eeafeded87 [ 959.601090] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.601305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.601540] env[61947]: DEBUG nova.objects.instance [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'resources' on Instance uuid ab027a9d-f51d-4cb6-8aaf-ebd507548657 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.636505] env[61947]: INFO nova.compute.manager [-] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Took 1.34 seconds to deallocate network for instance. [ 959.749514] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7aac5e13-a46c-499b-8d01-821ea4a67e8a tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.743s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.783762] env[61947]: DEBUG nova.objects.base [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 959.784713] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbb3e53-bfd1-4c47-a6ab-7e0edd3d0b01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.805622] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7ee5a0d-0382-41fc-a9f1-d7bf943591a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.811656] env[61947]: DEBUG oslo_vmware.api [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 959.811656] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5250354e-aa09-53a0-6b54-97463ff650ca" [ 959.811656] env[61947]: _type = "Task" [ 959.811656] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.819338] env[61947]: DEBUG oslo_vmware.api [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5250354e-aa09-53a0-6b54-97463ff650ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.061515] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0daf0b7-c314-4134-8fef-b18ef2ba26e9 tempest-ImagesTestJSON-357797237 tempest-ImagesTestJSON-357797237-project-member] Lock "e6435156-fb2e-4912-9587-99eeafeded87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.927s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.146842] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.234112] env[61947]: DEBUG nova.network.neutron [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.266293] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d577e7fc-4d02-45b7-bd0c-cb3654305682 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.274871] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c739c77-bd8e-4700-b1be-88b764e80188 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.309950] env[61947]: DEBUG nova.objects.instance [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.313171] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1c8b47-f59e-4633-a096-9d56a5eb4c29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.327139] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef0ada2-f925-4c3e-85cd-84e6fc263886 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.331790] env[61947]: DEBUG oslo_vmware.api [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5250354e-aa09-53a0-6b54-97463ff650ca, 'name': SearchDatastore_Task, 'duration_secs': 0.016716} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.332802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.345119] env[61947]: DEBUG nova.compute.provider_tree [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.420287] env[61947]: DEBUG nova.compute.manager [req-12cfd477-2168-4b4a-8a45-56bc631dc8b3 req-0d10101a-3c27-4dce-8849-5d4ef1206ee3 service nova] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Received event network-vif-deleted-e75bf013-ccbc-4dc2-ad7f-efa662784f0c {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.420491] env[61947]: DEBUG nova.compute.manager [req-12cfd477-2168-4b4a-8a45-56bc631dc8b3 req-0d10101a-3c27-4dce-8849-5d4ef1206ee3 service nova] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Received event network-vif-deleted-013a9067-5135-456a-812a-791fd1e20a3b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.616475] env[61947]: DEBUG nova.compute.manager [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.616706] env[61947]: DEBUG nova.compute.manager [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 960.617547] env[61947]: DEBUG oslo_concurrency.lockutils [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.617547] env[61947]: DEBUG oslo_concurrency.lockutils [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.617547] env[61947]: DEBUG nova.network.neutron [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.736561] env[61947]: INFO nova.compute.manager [-] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Took 1.26 seconds to deallocate network for instance. [ 960.819096] env[61947]: DEBUG oslo_concurrency.lockutils [None req-36702a3f-aa18-4ccb-ae69-735ee2ac9917 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.300s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.847730] env[61947]: DEBUG nova.scheduler.client.report [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.913133] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.913441] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.913637] env[61947]: DEBUG nova.compute.manager [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.914589] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b6ce01-4f22-459b-8cc5-168e23fdd4ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.922774] env[61947]: DEBUG nova.compute.manager [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 960.923371] env[61947]: DEBUG nova.objects.instance [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.243019] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.353050] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.355414] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.209s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.355639] env[61947]: DEBUG nova.objects.instance [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'resources' on Instance uuid ab21547e-0ddb-4271-967f-bea5056c163a {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.372597] env[61947]: INFO nova.scheduler.client.report [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted allocations for instance ab027a9d-f51d-4cb6-8aaf-ebd507548657 [ 961.428677] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 961.428978] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06a751d1-d3b5-4556-b790-ee295fe8bb18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.436839] env[61947]: DEBUG oslo_vmware.api [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 961.436839] env[61947]: value = "task-1224756" [ 961.436839] env[61947]: _type = "Task" [ 961.436839] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.449611] env[61947]: DEBUG oslo_vmware.api [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.501575] env[61947]: DEBUG nova.network.neutron [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.501965] env[61947]: DEBUG nova.network.neutron [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.696526] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.696864] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.882053] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8a5f51e4-878a-46cc-98ea-ff4d44091a4d tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "ab027a9d-f51d-4cb6-8aaf-ebd507548657" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.375s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.947440] env[61947]: DEBUG oslo_vmware.api [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224756, 'name': PowerOffVM_Task, 'duration_secs': 0.270552} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.947817] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 961.948527] env[61947]: DEBUG nova.compute.manager [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.951113] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456b4b7a-86b1-4d30-bac9-ef8b232f135e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.005026] env[61947]: DEBUG oslo_concurrency.lockutils [req-840f8b97-826a-4b9c-9925-5e994c867a35 req-d4c1c2f7-bbe0-4434-9539-61e472a7ef34 service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.054341] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dd2ecc-9b38-4972-a450-1db4204dfa3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.062186] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e613f8b-6d38-42e4-b815-4976c3ac0b48 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.092990] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07af2735-9253-40dd-a126-2c76774808cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.101411] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffe64b1-91ee-4612-8906-4b442455ba2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.122097] env[61947]: DEBUG nova.compute.provider_tree [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.199410] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.468406] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3352ade-3e3a-4601-b827-997a311c99da tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.626014] env[61947]: DEBUG nova.scheduler.client.report [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.683369] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.683369] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.723199] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.130075] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.132408] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.800s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.166573] env[61947]: INFO nova.scheduler.client.report [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance ab21547e-0ddb-4271-967f-bea5056c163a [ 963.186340] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.678324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fd403224-56de-4aa8-b5c9-b28c5985552c tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "ab21547e-0ddb-4271-967f-bea5056c163a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.485s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.713639] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.722881] env[61947]: DEBUG nova.objects.instance [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.815750] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab27a4c-7085-40f3-93f5-130ca2ed4668 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.824130] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f045872-0604-4fa0-9b7f-0c2aa29d773a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.855570] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdc22fe-5aca-48cc-a18a-250607a1f085 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.863133] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779830ac-497b-4e96-9f9e-099d6504b835 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.876733] env[61947]: DEBUG nova.compute.provider_tree [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.227742] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.228737] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.228737] env[61947]: DEBUG nova.network.neutron [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.228737] env[61947]: DEBUG nova.objects.instance [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'info_cache' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.379393] env[61947]: DEBUG nova.scheduler.client.report [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.426044] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.426044] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.731620] env[61947]: DEBUG nova.objects.base [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Object Instance<21bb0270-bc20-4ec1-9599-d676845b0dc7> lazy-loaded attributes: flavor,info_cache {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 964.928786] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 965.393079] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.261s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.395938] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.153s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.396180] env[61947]: DEBUG nova.objects.instance [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lazy-loading 'resources' on Instance uuid d2d094c4-0237-48cd-b98b-8d6c49ee5d94 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.460798] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.546549] env[61947]: DEBUG nova.network.neutron [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.660648] env[61947]: INFO nova.compute.manager [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Rebuilding instance [ 965.663411] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.663660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.663875] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.664066] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.664249] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.669076] env[61947]: INFO nova.compute.manager [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Terminating instance [ 965.670034] env[61947]: DEBUG nova.compute.manager [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.670328] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 965.671091] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddfe129-7b15-497e-91ac-3933f6017fff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.679301] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.681314] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40b3efdf-ff01-4f88-a948-3a6161cbd98c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.688499] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 965.688499] env[61947]: value = "task-1224757" [ 965.688499] env[61947]: _type = "Task" [ 965.688499] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.705299] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.712345] env[61947]: DEBUG nova.compute.manager [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.713247] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09503cb2-a050-469f-863a-182af8927a2c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.973150] env[61947]: INFO nova.scheduler.client.report [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocation for migration 6a038d9b-8359-4af1-a990-eab5cb334626 [ 966.049331] env[61947]: DEBUG oslo_concurrency.lockutils [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.084609] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036d0559-6cba-4aaf-891e-1c708e284222 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.092659] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f452c02-c8fc-4bd6-ac86-766e40436e4d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.124433] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebff469-3e8f-4925-a64e-9d3b8de0bd42 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.132908] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa4b1e5-0d58-44fa-a951-433ff420e35f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.147615] env[61947]: DEBUG nova.compute.provider_tree [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.199938] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224757, 'name': PowerOffVM_Task, 'duration_secs': 0.238338} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.200373] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.200590] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.200896] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c15b5df8-cac3-4ffc-8670-5120da5ed73c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.223617] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.224336] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-facd7cc8-3f05-47fb-b19b-3adb84d2b19d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.231670] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 966.231670] env[61947]: value = "task-1224759" [ 966.231670] env[61947]: _type = "Task" [ 966.231670] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.240099] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224759, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.274377] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.274876] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.275194] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleting the datastore file [datastore1] fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.275513] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-299807e9-b678-41a1-b0d8-89dba7568b6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.282572] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for the task: (returnval){ [ 966.282572] env[61947]: value = "task-1224760" [ 966.282572] env[61947]: _type = "Task" [ 966.282572] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.290386] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.404374] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.404666] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.481752] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dc8a1945-355b-48ea-ba79-87ed79dfffd7 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.580s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.554263] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.554694] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2689b5f-9108-4df3-ae9f-14785785fade {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.565080] env[61947]: DEBUG oslo_vmware.api [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 966.565080] env[61947]: value = "task-1224761" [ 966.565080] env[61947]: _type = "Task" [ 966.565080] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.573390] env[61947]: DEBUG oslo_vmware.api [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.650944] env[61947]: DEBUG nova.scheduler.client.report [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.745390] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.746068] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.746448] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.746798] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.747099] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.752022] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224759, 'name': PowerOffVM_Task, 'duration_secs': 0.182293} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.752022] env[61947]: INFO nova.compute.manager [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Terminating instance [ 966.752628] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.753450] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.754165] env[61947]: DEBUG nova.compute.manager [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.754451] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 966.754781] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebf4fe0e-bc31-4691-92d9-40ad8f7b8ad3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.757295] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d994069e-cfb7-4eb6-aa83-3197f86a0b3f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.765910] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.767388] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a52122e6-bbbd-4790-acd7-c50edbf7ef9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.769230] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 966.769230] env[61947]: value = "task-1224762" [ 966.769230] env[61947]: _type = "Task" [ 966.769230] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.775044] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 966.775044] env[61947]: value = "task-1224763" [ 966.775044] env[61947]: _type = "Task" [ 966.775044] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.783220] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 966.783220] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 966.783220] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264671', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'name': 'volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '91205e54-6bcb-403c-8308-0f0692fe4fba', 'attached_at': '', 'detached_at': '', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'serial': '4639dbb5-6722-408e-92c6-d142b4cdeb3b'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 966.783916] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ece988-42fc-4855-ba7b-c27a950d1592 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.793511] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.814670] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1216e71b-bec9-47fb-80ab-675856ff2369 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.817908] env[61947]: DEBUG oslo_vmware.api [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Task: {'id': task-1224760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160333} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.818287] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.818558] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 966.818835] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 966.819090] env[61947]: INFO nova.compute.manager [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 966.819583] env[61947]: DEBUG oslo.service.loopingcall [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.820140] env[61947]: DEBUG nova.compute.manager [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.820497] env[61947]: DEBUG nova.network.neutron [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 966.824894] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e126a47-580d-4142-a8fe-96072a762f78 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.845073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfa34e6-b51e-4ad1-addb-57ef648ca230 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.862733] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] The volume has not been displaced from its original location: [datastore1] volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b/volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 966.867658] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Reconfiguring VM instance instance-0000004c to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 966.868131] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3de90b47-8f0f-41a6-968b-7b5489d76f8a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.888306] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 966.888306] env[61947]: value = "task-1224764" [ 966.888306] env[61947]: _type = "Task" [ 966.888306] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.899360] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.907756] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.075581] env[61947]: DEBUG oslo_vmware.api [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224761, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.162336] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.167091] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.444s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.170373] env[61947]: INFO nova.compute.claims [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.211513] env[61947]: INFO nova.scheduler.client.report [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleted allocations for instance d2d094c4-0237-48cd-b98b-8d6c49ee5d94 [ 967.284890] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224763, 'name': PowerOffVM_Task, 'duration_secs': 0.166301} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.286432] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 967.286607] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 967.289452] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aea960cb-bcfb-4c13-9060-3527a0a6fc00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.291352] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "41f62ad6-b11e-4c77-98bc-47e26234112c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.291577] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.345907] env[61947]: DEBUG nova.compute.manager [req-ffb1c82f-d28e-4655-847c-a27d5b3bb6e1 req-5bc052b5-7f11-4717-8ef7-68070083cc38 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Received event network-vif-deleted-f6236fb9-a6fe-4b8d-948a-0518bada02e3 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.346189] env[61947]: INFO nova.compute.manager [req-ffb1c82f-d28e-4655-847c-a27d5b3bb6e1 req-5bc052b5-7f11-4717-8ef7-68070083cc38 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Neutron deleted interface f6236fb9-a6fe-4b8d-948a-0518bada02e3; detaching it from the instance and deleting it from the info cache [ 967.346374] env[61947]: DEBUG nova.network.neutron [req-ffb1c82f-d28e-4655-847c-a27d5b3bb6e1 req-5bc052b5-7f11-4717-8ef7-68070083cc38 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.384902] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 967.385159] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 967.385342] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleting the datastore file [datastore2] dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.385598] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66bff315-dacf-43f1-8611-fed28774de81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.393616] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 967.393616] env[61947]: value = "task-1224766" [ 967.393616] env[61947]: _type = "Task" [ 967.393616] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.400440] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224764, 'name': ReconfigVM_Task, 'duration_secs': 0.163393} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.400601] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Reconfigured VM instance instance-0000004c to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 967.407974] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d0d2457-d495-4d47-a780-f576ffcc4f5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.418499] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.427355] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 967.427355] env[61947]: value = "task-1224767" [ 967.427355] env[61947]: _type = "Task" [ 967.427355] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.436863] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224767, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.437997] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.576197] env[61947]: DEBUG oslo_vmware.api [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224761, 'name': PowerOnVM_Task, 'duration_secs': 0.623503} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.576197] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.576344] env[61947]: DEBUG nova.compute.manager [None req-bae2bd97-a5e6-45ec-b649-ccfddb28094e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.577369] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b75e45-7c76-4867-99f3-3f3003883aad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.727123] env[61947]: DEBUG oslo_concurrency.lockutils [None req-6d1fa0e8-32a0-4fa1-9ca1-43430179a94d tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "d2d094c4-0237-48cd-b98b-8d6c49ee5d94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.399s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.794569] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.817589] env[61947]: DEBUG nova.network.neutron [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.850170] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12b4b28a-166e-413c-8702-394eb90d4ae5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.858376] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cab7ff-88a3-4862-a67c-647af486b98c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.885298] env[61947]: DEBUG nova.compute.manager [req-ffb1c82f-d28e-4655-847c-a27d5b3bb6e1 req-5bc052b5-7f11-4717-8ef7-68070083cc38 service nova] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Detach interface failed, port_id=f6236fb9-a6fe-4b8d-948a-0518bada02e3, reason: Instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 967.903722] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.940035] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224767, 'name': ReconfigVM_Task, 'duration_secs': 0.309563} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.940153] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264671', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'name': 'volume-4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '91205e54-6bcb-403c-8308-0f0692fe4fba', 'attached_at': '', 'detached_at': '', 'volume_id': '4639dbb5-6722-408e-92c6-d142b4cdeb3b', 'serial': '4639dbb5-6722-408e-92c6-d142b4cdeb3b'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 967.940446] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 967.941554] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cb52b7-1951-473f-9fe8-25c3055491a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.950492] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 967.950880] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-215a2791-b6b3-472e-99e9-8fff10f9f66b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.011470] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.011738] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.011936] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Deleting the datastore file [datastore1] 91205e54-6bcb-403c-8308-0f0692fe4fba {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.012221] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-682c53d7-8053-45dd-b503-25419424b716 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.023169] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for the task: (returnval){ [ 968.023169] env[61947]: value = "task-1224770" [ 968.023169] env[61947]: _type = "Task" [ 968.023169] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.030749] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.311344] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.319890] env[61947]: INFO nova.compute.manager [-] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Took 1.50 seconds to deallocate network for instance. [ 968.339506] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d513c681-4b54-41c6-9d3a-b8f2ecc2a05c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.347314] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df89d1c-87f1-4a0f-9bce-9ff310246eaf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.380623] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608bbe9c-e4e6-4110-b1f0-c9ee0f5fbb49 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.388062] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f01ba68-7200-49f9-ac43-397e4207545b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.401444] env[61947]: DEBUG nova.compute.provider_tree [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.410188] env[61947]: DEBUG oslo_vmware.api [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224766, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.545742} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.410977] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.411188] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.411448] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.411531] env[61947]: INFO nova.compute.manager [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Took 1.66 seconds to destroy the instance on the hypervisor. [ 968.411756] env[61947]: DEBUG oslo.service.loopingcall [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.411941] env[61947]: DEBUG nova.compute.manager [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.412050] env[61947]: DEBUG nova.network.neutron [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 968.533066] env[61947]: DEBUG oslo_vmware.api [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Task: {'id': task-1224770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.069058} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.533343] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.533533] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.533714] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.590962] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 968.591581] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee51b0ba-62f2-4902-9fdb-38ce0592b436 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.600876] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1d8d0b-fd14-41fa-ac59-36754d8b7b3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.627908] env[61947]: ERROR nova.compute.manager [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Failed to detach volume 4639dbb5-6722-408e-92c6-d142b4cdeb3b from /dev/sda: nova.exception.InstanceNotFound: Instance 91205e54-6bcb-403c-8308-0f0692fe4fba could not be found. [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Traceback (most recent call last): [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self.driver.rebuild(**kwargs) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise NotImplementedError() [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] NotImplementedError [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] During handling of the above exception, another exception occurred: [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Traceback (most recent call last): [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self.driver.detach_volume(context, old_connection_info, [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] return self._volumeops.detach_volume(connection_info, instance) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._detach_volume_vmdk(connection_info, instance) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] stable_ref.fetch_moref(session) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise exception.InstanceNotFound(instance_id=self._uuid) [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] nova.exception.InstanceNotFound: Instance 91205e54-6bcb-403c-8308-0f0692fe4fba could not be found. [ 968.627908] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.756357] env[61947]: DEBUG nova.compute.utils [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Build of instance 91205e54-6bcb-403c-8308-0f0692fe4fba aborted: Failed to rebuild volume backed instance. {{(pid=61947) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 968.759287] env[61947]: ERROR nova.compute.manager [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 91205e54-6bcb-403c-8308-0f0692fe4fba aborted: Failed to rebuild volume backed instance. [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Traceback (most recent call last): [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self.driver.rebuild(**kwargs) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise NotImplementedError() [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] NotImplementedError [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] During handling of the above exception, another exception occurred: [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Traceback (most recent call last): [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._detach_root_volume(context, instance, root_bdm) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] with excutils.save_and_reraise_exception(): [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self.force_reraise() [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise self.value [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self.driver.detach_volume(context, old_connection_info, [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] return self._volumeops.detach_volume(connection_info, instance) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._detach_volume_vmdk(connection_info, instance) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] stable_ref.fetch_moref(session) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise exception.InstanceNotFound(instance_id=self._uuid) [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] nova.exception.InstanceNotFound: Instance 91205e54-6bcb-403c-8308-0f0692fe4fba could not be found. [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] During handling of the above exception, another exception occurred: [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Traceback (most recent call last): [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] yield [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 968.759287] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._do_rebuild_instance_with_claim( [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._do_rebuild_instance( [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._rebuild_default_impl(**kwargs) [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] self._rebuild_volume_backed_instance( [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] raise exception.BuildAbortException( [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] nova.exception.BuildAbortException: Build of instance 91205e54-6bcb-403c-8308-0f0692fe4fba aborted: Failed to rebuild volume backed instance. [ 968.760749] env[61947]: ERROR nova.compute.manager [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] [ 968.829075] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.904085] env[61947]: DEBUG nova.scheduler.client.report [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.132953] env[61947]: DEBUG nova.network.neutron [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.371061] env[61947]: DEBUG nova.compute.manager [req-77308f74-aa71-4e96-ae71-57c64fe17db4 req-7011b14d-38a7-48fe-966d-9df69db3dd97 service nova] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Received event network-vif-deleted-3adc00d1-4b5f-4857-a2ae-bdc4201f8c69 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.408234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.408737] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.411377] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.698s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.412817] env[61947]: INFO nova.compute.claims [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.635771] env[61947]: INFO nova.compute.manager [-] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Took 1.22 seconds to deallocate network for instance. [ 969.918596] env[61947]: DEBUG nova.compute.utils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.922307] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 969.922489] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 969.971900] env[61947]: DEBUG nova.policy [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 970.142400] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.286208] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Successfully created port: 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.421866] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.573102] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c47d044-36fe-4e7a-a0f2-a338d954044a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.580642] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed101f33-50f7-4878-a9a0-5ef7d66b0d58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.609434] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5553af6a-b8e2-4725-b594-b56195bec42d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.616019] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866baab3-d083-40c5-a383-0e628ba38cfe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.630173] env[61947]: DEBUG nova.compute.provider_tree [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.775449] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.133517] env[61947]: DEBUG nova.scheduler.client.report [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.431818] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.456194] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.456466] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.456627] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.456826] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.456989] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.457157] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.457367] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.457529] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.457699] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.457866] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.458062] env[61947]: DEBUG nova.virt.hardware [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.458936] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3473e92-a539-478e-b941-36a2401d7f81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.466868] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fc3593-dfa5-4079-a23c-2a814fc23fc4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.599308] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.599588] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.599889] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.600118] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.600337] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.603193] env[61947]: INFO nova.compute.manager [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Terminating instance [ 971.605009] env[61947]: DEBUG nova.compute.manager [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.605262] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.606264] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5e6ea1-de6f-4093-bb0e-c7648af07cc2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.614456] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.614715] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e95ad62-d042-4a14-89d4-5d39175cb065 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.621578] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 971.621578] env[61947]: value = "task-1224771" [ 971.621578] env[61947]: _type = "Task" [ 971.621578] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.629856] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.638333] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.227s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.638939] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 971.641874] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.181s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.644159] env[61947]: INFO nova.compute.claims [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.759392] env[61947]: DEBUG nova.compute.manager [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-vif-plugged-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.759808] env[61947]: DEBUG oslo_concurrency.lockutils [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.760202] env[61947]: DEBUG oslo_concurrency.lockutils [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.760488] env[61947]: DEBUG oslo_concurrency.lockutils [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.760791] env[61947]: DEBUG nova.compute.manager [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] No waiting events found dispatching network-vif-plugged-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.761037] env[61947]: WARNING nova.compute.manager [req-f3cd8432-61ab-4242-9679-6379cb533a8f req-fbaba7ba-3f5f-4967-a23a-476ecca2b109 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received unexpected event network-vif-plugged-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 for instance with vm_state building and task_state spawning. [ 971.864943] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Successfully updated port: 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.069805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "91205e54-6bcb-403c-8308-0f0692fe4fba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.069805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.069805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.069805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.069805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.071721] env[61947]: INFO nova.compute.manager [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Terminating instance [ 972.072657] env[61947]: DEBUG nova.compute.manager [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.072943] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8138b664-34eb-457b-b4bf-d0fb6c4a202a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.081979] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a80deb-71a8-42df-8534-11423fdaba86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.106661] env[61947]: WARNING nova.virt.vmwareapi.driver [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 91205e54-6bcb-403c-8308-0f0692fe4fba could not be found. [ 972.106871] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.107173] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb558725-6a63-4fc4-bf76-760c1dd44e71 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.114125] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfa458a-5d20-400b-bf97-5c37003e3bcc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.130586] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224771, 'name': PowerOffVM_Task, 'duration_secs': 0.184582} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.130805] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.130978] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.131304] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d47a1ec9-88f6-46f0-881b-3924e1d82236 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.140402] env[61947]: WARNING nova.virt.vmwareapi.vmops [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91205e54-6bcb-403c-8308-0f0692fe4fba could not be found. [ 972.140594] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.140775] env[61947]: INFO nova.compute.manager [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Took 0.07 seconds to destroy the instance on the hypervisor. [ 972.141016] env[61947]: DEBUG oslo.service.loopingcall [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.141347] env[61947]: DEBUG nova.compute.manager [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.141462] env[61947]: DEBUG nova.network.neutron [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.148547] env[61947]: DEBUG nova.compute.utils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.149971] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 972.150115] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 972.216909] env[61947]: DEBUG nova.policy [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '382358439d2c4512b1450ed16bf277bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5273b44dfca848538176692ab2a19e3d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 972.246340] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.246597] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.246834] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.247089] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.247211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.249322] env[61947]: INFO nova.compute.manager [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Terminating instance [ 972.251223] env[61947]: DEBUG nova.compute.manager [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.251416] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.252321] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b3a943-db68-4bf8-a28a-680ca546f736 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.260244] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.260472] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38e31fdb-badd-42b8-b9d4-a86fce22e985 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.266452] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 972.266452] env[61947]: value = "task-1224773" [ 972.266452] env[61947]: _type = "Task" [ 972.266452] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.273768] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.289036] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.289204] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.289389] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleting the datastore file [datastore2] 8c342d64-75f6-4b7d-829d-889eb3b6122f {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.289720] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e61fc637-5fd9-4a42-8543-d55291818c0f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.295998] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for the task: (returnval){ [ 972.295998] env[61947]: value = "task-1224774" [ 972.295998] env[61947]: _type = "Task" [ 972.295998] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.305904] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.367943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.368120] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.368526] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.655161] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 972.776085] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224773, 'name': PowerOffVM_Task, 'duration_secs': 0.168371} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.779118] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Successfully created port: 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.781126] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.781335] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.781788] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53885adf-093c-49e2-b475-435c3a1c5c38 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.805643] env[61947]: DEBUG oslo_vmware.api [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Task: {'id': task-1224774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139488} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.808497] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.808723] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.808965] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.809163] env[61947]: INFO nova.compute.manager [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Took 1.20 seconds to destroy the instance on the hypervisor. [ 972.809409] env[61947]: DEBUG oslo.service.loopingcall [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.809799] env[61947]: DEBUG nova.compute.manager [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.809903] env[61947]: DEBUG nova.network.neutron [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.842268] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.842510] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.842698] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleting the datastore file [datastore2] c4647324-7e52-44d4-a4b0-961684cc1c4c {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.845352] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3723bcf8-56aa-480e-b06b-295cf52da658 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.852891] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 972.852891] env[61947]: value = "task-1224776" [ 972.852891] env[61947]: _type = "Task" [ 972.852891] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.865397] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224776, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.874174] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464c8abe-643c-4a79-91a4-be3f09976a1e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.881883] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337a6d24-a9f7-4406-971e-2df52ae5d939 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.917411] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca102707-6381-44a3-b457-1ba8207d7002 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.929279] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f04675a-b5ec-4c7c-a279-6f671272c0cc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.939153] env[61947]: DEBUG nova.compute.provider_tree [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.940989] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.275365] env[61947]: DEBUG nova.network.neutron [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.363648] env[61947]: DEBUG oslo_vmware.api [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224776, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13878} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.363995] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.364100] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.364265] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.364440] env[61947]: INFO nova.compute.manager [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 973.364676] env[61947]: DEBUG oslo.service.loopingcall [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.364864] env[61947]: DEBUG nova.compute.manager [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.364957] env[61947]: DEBUG nova.network.neutron [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 973.477612] env[61947]: DEBUG nova.network.neutron [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.481832] env[61947]: ERROR nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [req-8efed8d8-7209-4195-866a-67595d922ee4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8efed8d8-7209-4195-866a-67595d922ee4"}]} [ 973.499604] env[61947]: DEBUG nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 973.515816] env[61947]: DEBUG nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 973.516080] env[61947]: DEBUG nova.compute.provider_tree [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.528650] env[61947]: DEBUG nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 973.548335] env[61947]: DEBUG nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 973.670021] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 973.709205] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.709469] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.709852] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.710214] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.710400] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.710557] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.710770] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.711170] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.711383] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.711558] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.714152] env[61947]: DEBUG nova.virt.hardware [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.714152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af72e7b-d207-4299-8d91-5f9d1e569505 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.724654] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3104ec63-b855-41ac-afc8-8b7b35fc98a7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.741710] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d5b59b-9933-4508-935c-06c6e851ed95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.748906] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77b8f1a-e8f5-4565-b2d6-3be0fc090800 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.778217] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.778544] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Instance network_info: |[{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.779203] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:53:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e4942ef-a8cf-44ec-bc80-58f9ae5894a5', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.786493] env[61947]: DEBUG oslo.service.loopingcall [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.787283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cf2e0f-8a2b-45ba-a604-9f02bf2820c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.791986] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 973.793118] env[61947]: DEBUG nova.compute.manager [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 973.794829] env[61947]: DEBUG nova.compute.manager [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing instance network info cache due to event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 973.795083] env[61947]: DEBUG oslo_concurrency.lockutils [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.795371] env[61947]: DEBUG oslo_concurrency.lockutils [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.795442] env[61947]: DEBUG nova.network.neutron [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.796525] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efdfd558-f92b-4bba-b4b5-bda9c6bd6499 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.819605] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c155ccf-6d6d-4026-8c0d-67fa0857d99d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.824666] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.824666] env[61947]: value = "task-1224777" [ 973.824666] env[61947]: _type = "Task" [ 973.824666] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.836113] env[61947]: DEBUG nova.compute.provider_tree [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.842141] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.984705] env[61947]: INFO nova.compute.manager [-] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Took 1.84 seconds to deallocate network for instance. [ 974.034098] env[61947]: DEBUG nova.network.neutron [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updated VIF entry in instance network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.034644] env[61947]: DEBUG nova.network.neutron [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.319975] env[61947]: DEBUG nova.network.neutron [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.336241] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.387673] env[61947]: DEBUG nova.scheduler.client.report [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 974.387958] env[61947]: DEBUG nova.compute.provider_tree [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 115 to 116 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.388160] env[61947]: DEBUG nova.compute.provider_tree [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.453690] env[61947]: DEBUG nova.compute.manager [req-c058e6d4-8432-4a53-a744-74968a9be9e1 req-cd0537d1-e8a1-4cc0-9003-09db55078e3d service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Received event network-vif-deleted-c70b72e2-fda2-45c5-afe8-1af9c339fdb2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.453920] env[61947]: INFO nova.compute.manager [req-c058e6d4-8432-4a53-a744-74968a9be9e1 req-cd0537d1-e8a1-4cc0-9003-09db55078e3d service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Neutron deleted interface c70b72e2-fda2-45c5-afe8-1af9c339fdb2; detaching it from the instance and deleting it from the info cache [ 974.454136] env[61947]: DEBUG nova.network.neutron [req-c058e6d4-8432-4a53-a744-74968a9be9e1 req-cd0537d1-e8a1-4cc0-9003-09db55078e3d service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.537742] env[61947]: DEBUG oslo_concurrency.lockutils [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.538050] env[61947]: DEBUG nova.compute.manager [req-08bc6429-fdde-441e-a4cb-28701e8594ac req-d527d170-f0ba-4d8f-8eab-fbb83a94832c service nova] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Received event network-vif-deleted-c2f0d861-74c8-4b27-abb4-94bfa51b7768 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.545674] env[61947]: INFO nova.compute.manager [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Took 0.56 seconds to detach 1 volumes for instance. [ 974.554221] env[61947]: DEBUG nova.compute.manager [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Deleting volume: 4639dbb5-6722-408e-92c6-d142b4cdeb3b {{(pid=61947) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 974.822104] env[61947]: INFO nova.compute.manager [-] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Took 2.01 seconds to deallocate network for instance. [ 974.836834] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.893866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.252s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.896029] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.897175] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.459s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.898660] env[61947]: INFO nova.compute.claims [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.933018] env[61947]: DEBUG nova.network.neutron [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.957209] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d0ae60a-6bd6-42a0-a7c5-b90dfecf0845 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.965293] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Successfully updated port: 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.969088] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58078f40-73f6-4cc9-970b-95507727dbcd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.997926] env[61947]: DEBUG nova.compute.manager [req-c058e6d4-8432-4a53-a744-74968a9be9e1 req-cd0537d1-e8a1-4cc0-9003-09db55078e3d service nova] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Detach interface failed, port_id=c70b72e2-fda2-45c5-afe8-1af9c339fdb2, reason: Instance c4647324-7e52-44d4-a4b0-961684cc1c4c could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 975.117912] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.333457] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.339613] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.407619] env[61947]: DEBUG nova.compute.utils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.409314] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.409662] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.435208] env[61947]: INFO nova.compute.manager [-] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Took 2.07 seconds to deallocate network for instance. [ 975.465232] env[61947]: DEBUG nova.policy [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 975.467174] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.467281] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.467436] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.806411] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Successfully created port: 76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.824203] env[61947]: DEBUG nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Received event network-vif-deleted-1a026df8-fa69-439e-805b-5dad5c772246 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.824450] env[61947]: DEBUG nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Received event network-vif-plugged-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.824789] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.824899] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.825082] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.825195] env[61947]: DEBUG nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] No waiting events found dispatching network-vif-plugged-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.825343] env[61947]: WARNING nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Received unexpected event network-vif-plugged-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 for instance with vm_state building and task_state spawning. [ 975.825509] env[61947]: DEBUG nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Received event network-changed-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.825680] env[61947]: DEBUG nova.compute.manager [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Refreshing instance network info cache due to event network-changed-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 975.825878] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.837734] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.913758] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.942851] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.002247] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.081744] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1940ca-468c-431e-9d6d-be9ca839aff5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.090175] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039daeda-817e-44a2-b18b-786fcb0129ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.125590] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291be325-e2e9-4533-a5ea-e2d76cb7e68b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.133580] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66072bd4-2abe-41fb-9ff6-04faa1d7ba0e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.147991] env[61947]: DEBUG nova.compute.provider_tree [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.175099] env[61947]: DEBUG nova.network.neutron [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.337728] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224777, 'name': CreateVM_Task, 'duration_secs': 2.363211} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.337902] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 976.338608] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.338780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.339171] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 976.339435] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ef3a6b4-0ef6-4ef2-981d-332bce304569 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.344416] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 976.344416] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5274f2e0-c49f-622f-0612-a21afe745c70" [ 976.344416] env[61947]: _type = "Task" [ 976.344416] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.352654] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5274f2e0-c49f-622f-0612-a21afe745c70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.651730] env[61947]: DEBUG nova.scheduler.client.report [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.678171] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.678496] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Instance network_info: |[{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 976.678835] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.679039] env[61947]: DEBUG nova.network.neutron [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Refreshing network info cache for port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.680178] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:12:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ec3f44d-0a7d-45d1-8429-6e2eccb59e93', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.688727] env[61947]: DEBUG oslo.service.loopingcall [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.689964] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 976.690245] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c744241c-8d7d-4ebf-b32c-aa5e398f2723 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.712206] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.712206] env[61947]: value = "task-1224779" [ 976.712206] env[61947]: _type = "Task" [ 976.712206] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.721954] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224779, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.857028] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5274f2e0-c49f-622f-0612-a21afe745c70, 'name': SearchDatastore_Task, 'duration_secs': 0.009436} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.857028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.857028] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.857028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.857028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.857028] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.857028] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c603dca6-9849-406c-a64e-d195431c4fa7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.864819] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.865181] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 976.865977] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4037eeb1-4676-4eb4-b80d-0f52c2e14bba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.873958] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 976.873958] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc7660-65c0-d096-cdba-63794a215fea" [ 976.873958] env[61947]: _type = "Task" [ 976.873958] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.880851] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc7660-65c0-d096-cdba-63794a215fea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.923805] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.954338] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.954699] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.954939] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.955270] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.955493] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.955715] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.956030] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.956268] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.956500] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.956728] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.956974] env[61947]: DEBUG nova.virt.hardware [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.958172] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdbd17e-7896-4dfa-b637-b064ca46e6da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.967787] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5358f289-206e-4bad-b851-42dcfc2dbb20 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.160975] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.160975] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.166020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.852s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.166020] env[61947]: INFO nova.compute.claims [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.221915] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224779, 'name': CreateVM_Task, 'duration_secs': 0.315505} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.222098] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.222762] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.222936] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.223283] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.223550] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18a7c492-35fa-451f-b648-ef740972ad7c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.228315] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 977.228315] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5279c7fc-266a-f51a-3720-a7cdc87f9bb5" [ 977.228315] env[61947]: _type = "Task" [ 977.228315] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.237678] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5279c7fc-266a-f51a-3720-a7cdc87f9bb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.382362] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dc7660-65c0-d096-cdba-63794a215fea, 'name': SearchDatastore_Task, 'duration_secs': 0.009302} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.383312] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c2b8173-3bbb-4e94-bf0b-b9af059dabd0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.391698] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 977.391698] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525601f4-e33b-128f-0a64-4c839989e835" [ 977.391698] env[61947]: _type = "Task" [ 977.391698] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.399388] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525601f4-e33b-128f-0a64-4c839989e835, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.457187] env[61947]: DEBUG nova.network.neutron [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updated VIF entry in instance network info cache for port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.457660] env[61947]: DEBUG nova.network.neutron [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.460118] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Successfully updated port: 76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.670646] env[61947]: DEBUG nova.compute.utils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.674448] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.674615] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 977.711089] env[61947]: DEBUG nova.policy [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66d98dc3ada744719c44d4f14c6555a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b957556421474151ae57dfb61507d6cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 977.742182] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5279c7fc-266a-f51a-3720-a7cdc87f9bb5, 'name': SearchDatastore_Task, 'duration_secs': 0.015654} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.742517] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.742753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.743260] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.851792] env[61947]: DEBUG nova.compute.manager [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Received event network-vif-plugged-76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.852044] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Acquiring lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.852275] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.852456] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.852873] env[61947]: DEBUG nova.compute.manager [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] No waiting events found dispatching network-vif-plugged-76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.852873] env[61947]: WARNING nova.compute.manager [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Received unexpected event network-vif-plugged-76040c0b-35be-42ca-92e1-56f8ebc53495 for instance with vm_state building and task_state spawning. [ 977.852951] env[61947]: DEBUG nova.compute.manager [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Received event network-changed-76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.853218] env[61947]: DEBUG nova.compute.manager [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Refreshing instance network info cache due to event network-changed-76040c0b-35be-42ca-92e1-56f8ebc53495. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 977.853450] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Acquiring lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.853563] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Acquired lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.853720] env[61947]: DEBUG nova.network.neutron [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Refreshing network info cache for port 76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.903335] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]525601f4-e33b-128f-0a64-4c839989e835, 'name': SearchDatastore_Task, 'duration_secs': 0.017249} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.903625] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.903848] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] e137f21e-766d-4b20-9d92-5d1907e2baa3/e137f21e-766d-4b20-9d92-5d1907e2baa3.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 977.904137] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.904325] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.904546] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee4f09bf-36fc-42f8-a081-12f1104995ac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.906634] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1254e346-3fcc-4b60-9a38-e4b039f4abf4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.913481] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 977.913481] env[61947]: value = "task-1224780" [ 977.913481] env[61947]: _type = "Task" [ 977.913481] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.918069] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.918197] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.919929] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c829b45-f951-4ee0-9f0f-fdb4ce8c1b15 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.926740] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.930420] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 977.930420] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5235c901-967f-2f4a-0fb1-be1425c788f2" [ 977.930420] env[61947]: _type = "Task" [ 977.930420] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.939947] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5235c901-967f-2f4a-0fb1-be1425c788f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.961812] env[61947]: DEBUG oslo_concurrency.lockutils [req-eddfb406-20a7-4892-92e3-863f533f3bcb req-a5b3a347-085a-4fb7-b455-3e545bc2ca34 service nova] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.962670] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.012906] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Successfully created port: b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.175864] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.388114] env[61947]: DEBUG nova.network.neutron [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 978.420373] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae7e438-e3e4-4f7c-a1e6-bbb6d11fb4f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.430802] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76b11a4-c53a-440a-9f48-b41857d0f151 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.434488] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224780, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478711} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.436778] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] e137f21e-766d-4b20-9d92-5d1907e2baa3/e137f21e-766d-4b20-9d92-5d1907e2baa3.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 978.437683] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.441598] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d2fc6b3-d8a3-460a-b885-bf6f744fc3e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.475011] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f741c5e-1398-42c4-aad2-bb44aa0d577a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.477578] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 978.477578] env[61947]: value = "task-1224781" [ 978.477578] env[61947]: _type = "Task" [ 978.477578] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.477821] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5235c901-967f-2f4a-0fb1-be1425c788f2, 'name': SearchDatastore_Task, 'duration_secs': 0.008309} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.482127] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1941a5da-232e-4dac-a2ce-3cb9eeb156a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.486639] env[61947]: DEBUG nova.network.neutron [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.491020] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb086d0-0693-470a-883c-a04a1725957c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.498633] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224781, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.498967] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 978.498967] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dbf33d-7b3a-5224-3854-6ed485994e3a" [ 978.498967] env[61947]: _type = "Task" [ 978.498967] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.511972] env[61947]: DEBUG nova.compute.provider_tree [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.518348] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52dbf33d-7b3a-5224-3854-6ed485994e3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.518595] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.518867] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 978.519140] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8fadb8b-8e1e-4f43-8f02-1974e011d07f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.525081] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 978.525081] env[61947]: value = "task-1224782" [ 978.525081] env[61947]: _type = "Task" [ 978.525081] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.533403] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224782, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.988060] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05841} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.988060] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.988674] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61754d90-babc-4e50-b6b1-9859d0d6b3f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.003549] env[61947]: DEBUG oslo_concurrency.lockutils [req-b95547ba-b9f0-4b33-8b54-71d3c1c0b401 req-bcbf45dc-bbd2-4c95-aa52-31dd2e44160d service nova] Releasing lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.014027] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] e137f21e-766d-4b20-9d92-5d1907e2baa3/e137f21e-766d-4b20-9d92-5d1907e2baa3.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.014027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.014027] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.015788] env[61947]: DEBUG nova.scheduler.client.report [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.019032] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6d6131d-3856-49fb-89c0-cf4a817162e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.040919] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.875s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.041488] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.044718] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.216s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.044927] env[61947]: DEBUG nova.objects.instance [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lazy-loading 'resources' on Instance uuid fa0eed6e-831f-4e7a-9890-fc5a02609aa4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.058383] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224782, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436993} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.058940] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 979.059193] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.060888] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 979.060888] env[61947]: value = "task-1224783" [ 979.060888] env[61947]: _type = "Task" [ 979.060888] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.060888] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de1e3177-4f59-4ea4-a058-105a4c5c42e9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.073655] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224783, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.075415] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 979.075415] env[61947]: value = "task-1224784" [ 979.075415] env[61947]: _type = "Task" [ 979.075415] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.080877] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 979.091037] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224784, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.187118] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.212722] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.213173] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.213314] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.213536] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.213696] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.213851] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.214074] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.214247] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.214406] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.214572] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.214747] env[61947]: DEBUG nova.virt.hardware [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.215604] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eef07fb-3cb9-436f-8e4e-0e816288c342 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.226243] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab4bbdd-b0c4-4d3c-b623-6d506d866f2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.257089] env[61947]: DEBUG nova.network.neutron [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Updating instance_info_cache with network_info: [{"id": "76040c0b-35be-42ca-92e1-56f8ebc53495", "address": "fa:16:3e:10:cc:84", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76040c0b-35", "ovs_interfaceid": "76040c0b-35be-42ca-92e1-56f8ebc53495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.459778] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.460864] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.549528] env[61947]: DEBUG nova.compute.utils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.554020] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.554020] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.575344] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224783, 'name': ReconfigVM_Task, 'duration_secs': 0.290622} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.575575] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfigured VM instance instance-00000055 to attach disk [datastore2] e137f21e-766d-4b20-9d92-5d1907e2baa3/e137f21e-766d-4b20-9d92-5d1907e2baa3.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.578900] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6af5308d-e437-4fc2-9be8-0593934c788d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.589636] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224784, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075091} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.591149] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.591566] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 979.591566] env[61947]: value = "task-1224785" [ 979.591566] env[61947]: _type = "Task" [ 979.591566] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.592290] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f39782-16c0-4ac8-8dae-e6aa9d289186 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.603349] env[61947]: DEBUG nova.policy [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0337baafe604194a5bc93e99ca8078a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51aa57efe97e453783044286f33d3f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 979.626149] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.626149] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224785, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.626900] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Successfully updated port: b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.628364] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c665aeb1-74bd-49a5-b437-a87687e5915e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.652532] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 979.652532] env[61947]: value = "task-1224786" [ 979.652532] env[61947]: _type = "Task" [ 979.652532] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.664865] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224786, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.760248] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-85210032-a4e2-4f76-96e7-efa2fa6e134e" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.760580] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance network_info: |[{"id": "76040c0b-35be-42ca-92e1-56f8ebc53495", "address": "fa:16:3e:10:cc:84", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76040c0b-35", "ovs_interfaceid": "76040c0b-35be-42ca-92e1-56f8ebc53495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.761137] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:cc:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76040c0b-35be-42ca-92e1-56f8ebc53495', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.770418] env[61947]: DEBUG oslo.service.loopingcall [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.773520] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 979.774880] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-449560f2-9916-4aa4-9368-f1e5bbfa4345 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.801198] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.801198] env[61947]: value = "task-1224787" [ 979.801198] env[61947]: _type = "Task" [ 979.801198] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.809662] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224787, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.822449] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afedb49-4007-4c86-a2a3-fd8160103c1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.829617] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e507c36-f231-4f9a-b5a9-5a60f56d67f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.860245] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59be48a3-db0d-41c8-a1cd-26d56164d8c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.867997] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4f3e13-43a5-4eee-9efc-b3268de88f1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.882641] env[61947]: DEBUG nova.compute.provider_tree [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.899867] env[61947]: DEBUG nova.compute.manager [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Received event network-vif-plugged-b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.900431] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Acquiring lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.900677] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.900861] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.901132] env[61947]: DEBUG nova.compute.manager [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] No waiting events found dispatching network-vif-plugged-b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 979.901345] env[61947]: WARNING nova.compute.manager [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Received unexpected event network-vif-plugged-b3e0634e-84a9-477c-b0fd-e027235d9cf4 for instance with vm_state building and task_state spawning. [ 979.901527] env[61947]: DEBUG nova.compute.manager [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Received event network-changed-b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.901683] env[61947]: DEBUG nova.compute.manager [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Refreshing instance network info cache due to event network-changed-b3e0634e-84a9-477c-b0fd-e027235d9cf4. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 979.901874] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Acquiring lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.902030] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Acquired lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.902195] env[61947]: DEBUG nova.network.neutron [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Refreshing network info cache for port b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.963830] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 980.057766] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.106586] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224785, 'name': Rename_Task, 'duration_secs': 0.141318} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.106995] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 980.107283] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7bac639c-9115-441d-921b-3946bc1e55f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.114427] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 980.114427] env[61947]: value = "task-1224788" [ 980.114427] env[61947]: _type = "Task" [ 980.114427] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.123309] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.147074] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.163132] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224786, 'name': ReconfigVM_Task, 'duration_secs': 0.405929} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.163461] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.164461] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57d9b3d4-f064-4071-bd2f-311beab75639 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.170965] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 980.170965] env[61947]: value = "task-1224789" [ 980.170965] env[61947]: _type = "Task" [ 980.170965] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.179481] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Successfully created port: bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.185041] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224789, 'name': Rename_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.311755] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224787, 'name': CreateVM_Task, 'duration_secs': 0.328032} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.311943] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 980.312766] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.312962] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.313350] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.313623] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a998f82-09e9-4f76-bda7-6a8eda03b4b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.318538] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 980.318538] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a67dcd-cd81-b891-9078-4436ab8222cf" [ 980.318538] env[61947]: _type = "Task" [ 980.318538] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.327383] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a67dcd-cd81-b891-9078-4436ab8222cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.415156] env[61947]: ERROR nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] [req-ebfe7ab2-78b3-4468-a28b-eb62a49cf2ec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ebfe7ab2-78b3-4468-a28b-eb62a49cf2ec"}]} [ 980.442598] env[61947]: DEBUG nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 980.452759] env[61947]: DEBUG nova.network.neutron [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.466501] env[61947]: DEBUG nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 980.466501] env[61947]: DEBUG nova.compute.provider_tree [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.483334] env[61947]: DEBUG nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 980.493943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.505675] env[61947]: DEBUG nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 980.582501] env[61947]: DEBUG nova.network.neutron [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.630242] env[61947]: DEBUG oslo_vmware.api [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224788, 'name': PowerOnVM_Task, 'duration_secs': 0.466025} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.630584] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 980.630767] env[61947]: INFO nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Took 9.20 seconds to spawn the instance on the hypervisor. [ 980.630922] env[61947]: DEBUG nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.631883] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009c7d28-6200-4bf1-a841-1a1c72c29d93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.683364] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224789, 'name': Rename_Task, 'duration_secs': 0.238748} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.683641] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 980.683890] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b101329f-51c0-4500-92aa-c2366cfe0431 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.691816] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 980.691816] env[61947]: value = "task-1224790" [ 980.691816] env[61947]: _type = "Task" [ 980.691816] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.699401] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.739902] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c29afe-81ef-4f99-8b5a-8784c43864a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.747144] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04f3c7d-a3d5-409e-833f-8a05c51584d7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.783131] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3aab991-bf51-4460-8a60-ad36988c1ff1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.790800] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a3307d-c4d9-4c5e-a63d-794194f1f2d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.806395] env[61947]: DEBUG nova.compute.provider_tree [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.827831] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a67dcd-cd81-b891-9078-4436ab8222cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009992} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.828141] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.828383] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.828618] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.828769] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.828984] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.829267] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fcb2a18e-373c-4737-9fb8-cfe2a35bf60d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.837047] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.837274] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 980.838093] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-113507d3-4e95-4262-a07e-72b0a16b82b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.843429] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 980.843429] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52592b59-9eb3-88de-7bc2-734e4826564a" [ 980.843429] env[61947]: _type = "Task" [ 980.843429] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.851720] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52592b59-9eb3-88de-7bc2-734e4826564a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.072973] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.084309] env[61947]: DEBUG oslo_concurrency.lockutils [req-b1b00015-f2a5-4580-b293-1ff26ab87375 req-8fdefe32-064f-4276-964e-190d11401f31 service nova] Releasing lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.084581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquired lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.085256] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.103205] env[61947]: DEBUG nova.virt.hardware [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.103800] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395ca216-2a31-453c-8b7b-ee770eaf0ba2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.111823] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867e456d-d28d-480c-bd03-5b4a24aba752 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.151166] env[61947]: INFO nova.compute.manager [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Took 18.44 seconds to build instance. [ 981.202377] env[61947]: DEBUG oslo_vmware.api [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224790, 'name': PowerOnVM_Task, 'duration_secs': 0.473485} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.202666] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 981.202869] env[61947]: INFO nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Took 7.53 seconds to spawn the instance on the hypervisor. [ 981.203069] env[61947]: DEBUG nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.204018] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a9a7dc-b8a6-45b6-819c-2be5bcddc82d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.337607] env[61947]: DEBUG nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 981.337984] env[61947]: DEBUG nova.compute.provider_tree [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 117 to 118 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 981.338123] env[61947]: DEBUG nova.compute.provider_tree [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.355770] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52592b59-9eb3-88de-7bc2-734e4826564a, 'name': SearchDatastore_Task, 'duration_secs': 0.008777} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.357418] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-794919f2-2676-4d31-8cd5-a018b8279a14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.363099] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 981.363099] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52707fc8-c290-279b-8894-070b3574ecd3" [ 981.363099] env[61947]: _type = "Task" [ 981.363099] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.371048] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52707fc8-c290-279b-8894-070b3574ecd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.640589] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.652681] env[61947]: DEBUG oslo_concurrency.lockutils [None req-79b22b7b-6d24-4de0-a1aa-eac532c965c3 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.956s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.704019] env[61947]: DEBUG nova.compute.manager [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Received event network-vif-plugged-bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.704169] env[61947]: DEBUG oslo_concurrency.lockutils [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] Acquiring lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.704380] env[61947]: DEBUG oslo_concurrency.lockutils [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.704554] env[61947]: DEBUG oslo_concurrency.lockutils [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.704771] env[61947]: DEBUG nova.compute.manager [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] No waiting events found dispatching network-vif-plugged-bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.704947] env[61947]: WARNING nova.compute.manager [req-90df9373-41fa-40a1-96c6-468c4f16b0ef req-d00d76d9-89b2-41d9-9b2b-24134e658d44 service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Received unexpected event network-vif-plugged-bb025d74-6765-4162-98a0-e008a39320fe for instance with vm_state building and task_state spawning. [ 981.728147] env[61947]: INFO nova.compute.manager [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Took 18.04 seconds to build instance. [ 981.848714] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.804s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.851662] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.709s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.851662] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.852819] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.078s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.876197] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52707fc8-c290-279b-8894-070b3574ecd3, 'name': SearchDatastore_Task, 'duration_secs': 0.008301} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.877154] env[61947]: INFO nova.scheduler.client.report [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Deleted allocations for instance fa0eed6e-831f-4e7a-9890-fc5a02609aa4 [ 981.878229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.878528] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 85210032-a4e2-4f76-96e7-efa2fa6e134e/85210032-a4e2-4f76-96e7-efa2fa6e134e.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 981.886540] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27b7bb3b-be50-4311-a986-91b7b8a41d6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.892211] env[61947]: INFO nova.scheduler.client.report [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocations for instance dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2 [ 981.902857] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 981.902857] env[61947]: value = "task-1224791" [ 981.902857] env[61947]: _type = "Task" [ 981.902857] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.909206] env[61947]: DEBUG nova.network.neutron [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Updating instance_info_cache with network_info: [{"id": "b3e0634e-84a9-477c-b0fd-e027235d9cf4", "address": "fa:16:3e:cd:cc:d4", "network": {"id": "154e3e24-4164-4d8c-938e-b73d81a32ecd", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2042061717-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b957556421474151ae57dfb61507d6cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3e0634e-84", "ovs_interfaceid": "b3e0634e-84a9-477c-b0fd-e027235d9cf4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.920082] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.924553] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Successfully updated port: bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.051985] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95782bda-2fc9-4696-95f9-a9cc6efc5694 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.062283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5d65d5-ee9c-43bc-9266-2bab2a3c297f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.099177] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf124d6b-c15d-4565-b3c9-43b3ac0835e0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.108695] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd194577-fae6-483a-b186-d0937f4c6678 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.118185] env[61947]: DEBUG nova.compute.manager [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Received event network-changed-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 982.118185] env[61947]: DEBUG nova.compute.manager [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Refreshing instance network info cache due to event network-changed-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 982.118454] env[61947]: DEBUG oslo_concurrency.lockutils [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.118539] env[61947]: DEBUG oslo_concurrency.lockutils [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.118701] env[61947]: DEBUG nova.network.neutron [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Refreshing network info cache for port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.129807] env[61947]: DEBUG nova.compute.provider_tree [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.230898] env[61947]: DEBUG oslo_concurrency.lockutils [None req-da59b135-f599-484a-8abe-f416b65c0b7e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.547s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.397437] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5c6e412d-6302-4650-aa53-8af3d058615d tempest-AttachVolumeShelveTestJSON-1973816620 tempest-AttachVolumeShelveTestJSON-1973816620-project-member] Lock "fa0eed6e-831f-4e7a-9890-fc5a02609aa4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.733s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.403321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c364e1e2-650c-4fc7-89ea-e32ba6c6e28a tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.657s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.411438] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Releasing lock "refresh_cache-ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.412624] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance network_info: |[{"id": "b3e0634e-84a9-477c-b0fd-e027235d9cf4", "address": "fa:16:3e:cd:cc:d4", "network": {"id": "154e3e24-4164-4d8c-938e-b73d81a32ecd", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2042061717-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b957556421474151ae57dfb61507d6cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3e0634e-84", "ovs_interfaceid": "b3e0634e-84a9-477c-b0fd-e027235d9cf4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.412624] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:cc:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf63c3c8-d774-4b81-9b12-848612a96076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3e0634e-84a9-477c-b0fd-e027235d9cf4', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.422750] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Creating folder: Project (b957556421474151ae57dfb61507d6cd). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 982.428421] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a03e129b-8945-4ae5-b78b-6b6b1b374a12 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.430599] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.430731] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.430869] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.435138] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48126} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.435138] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 85210032-a4e2-4f76-96e7-efa2fa6e134e/85210032-a4e2-4f76-96e7-efa2fa6e134e.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 982.435138] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.435138] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb9a7196-e528-411d-92b1-c4f627ef35ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.442291] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 982.442291] env[61947]: value = "task-1224793" [ 982.442291] env[61947]: _type = "Task" [ 982.442291] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.448218] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Created folder: Project (b957556421474151ae57dfb61507d6cd) in parent group-v264556. [ 982.448218] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Creating folder: Instances. Parent ref: group-v264710. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 982.448658] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34c359ca-09f5-4f44-9c7d-88efe053fd9a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.455129] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224793, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.468790] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Created folder: Instances in parent group-v264710. [ 982.468790] env[61947]: DEBUG oslo.service.loopingcall [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.468790] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 982.468790] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb28aef1-620f-4c5b-ac8c-9e5da2129e29 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.493278] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.493278] env[61947]: value = "task-1224795" [ 982.493278] env[61947]: _type = "Task" [ 982.493278] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.501454] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224795, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.636761] env[61947]: DEBUG nova.scheduler.client.report [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.953494] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224793, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069494} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.955955] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.956818] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b74dea-4383-4e0e-8c6d-84303c3fe169 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.987423] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 85210032-a4e2-4f76-96e7-efa2fa6e134e/85210032-a4e2-4f76-96e7-efa2fa6e134e.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.987819] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bc299d2-5b13-4046-830e-4f09cab9bb12 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.017923] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224795, 'name': CreateVM_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.019583] env[61947]: DEBUG nova.network.neutron [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updated VIF entry in instance network info cache for port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.019955] env[61947]: DEBUG nova.network.neutron [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.022749] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 983.022749] env[61947]: value = "task-1224796" [ 983.022749] env[61947]: _type = "Task" [ 983.022749] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.031694] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224796, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.122426] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 983.140593] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.287s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.141387] env[61947]: INFO nova.compute.manager [None req-9f29cce4-0efd-4111-addd-5eb9dbc40d8f tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Successfully reverted task state from rebuilding on failure for instance. [ 983.153719] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.036s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.154184] env[61947]: DEBUG nova.objects.instance [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lazy-loading 'resources' on Instance uuid 91205e54-6bcb-403c-8308-0f0692fe4fba {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.444013] env[61947]: DEBUG nova.network.neutron [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Updating instance_info_cache with network_info: [{"id": "bb025d74-6765-4162-98a0-e008a39320fe", "address": "fa:16:3e:a1:37:dc", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb025d74-67", "ovs_interfaceid": "bb025d74-6765-4162-98a0-e008a39320fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.522163] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224795, 'name': CreateVM_Task, 'duration_secs': 0.844319} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.522163] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 983.524172] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.524392] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.524736] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.525275] env[61947]: DEBUG oslo_concurrency.lockutils [req-d15206a7-8fbf-49bf-af8b-d13b03bc6149 req-afee4514-7714-46c5-b068-55eaa0edcb98 service nova] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.525698] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41efa867-319c-4c45-8d6c-19555b273f06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.534833] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 983.534833] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5299df38-76d9-512f-a857-e5e3b2778839" [ 983.534833] env[61947]: _type = "Task" [ 983.534833] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.538822] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224796, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.547849] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5299df38-76d9-512f-a857-e5e3b2778839, 'name': SearchDatastore_Task, 'duration_secs': 0.010603} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.548894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.548894] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.548894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.548894] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.549145] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.550047] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06102ef6-fb34-464b-a08a-6345e34aad9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.559391] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.559391] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 983.559391] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7664a84a-bfce-4303-97d7-9ceac1b9afe0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.565945] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 983.565945] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520d30ce-aae6-947c-fd08-3f0ee36dcf30" [ 983.565945] env[61947]: _type = "Task" [ 983.565945] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.578635] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520d30ce-aae6-947c-fd08-3f0ee36dcf30, 'name': SearchDatastore_Task, 'duration_secs': 0.00886} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.579792] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8fc6094-cf93-4b37-b649-8b83b1df45de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.591838] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 983.591838] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a541e7-4d6d-3f2c-cc2a-3917dfddd046" [ 983.591838] env[61947]: _type = "Task" [ 983.591838] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.606504] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a541e7-4d6d-3f2c-cc2a-3917dfddd046, 'name': SearchDatastore_Task, 'duration_secs': 0.009641} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.606504] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.606504] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ebf276db-cc9b-441f-a01b-7f7dc5b83fd3/ebf276db-cc9b-441f-a01b-7f7dc5b83fd3.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.606504] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d67e19d-f24b-4509-8598-0914652d173d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.613619] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 983.613619] env[61947]: value = "task-1224797" [ 983.613619] env[61947]: _type = "Task" [ 983.613619] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.622437] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224797, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.744779] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Received event network-changed-bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.745052] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Refreshing instance network info cache due to event network-changed-bb025d74-6765-4162-98a0-e008a39320fe. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 983.745729] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquiring lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.883511] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34b6535-66f9-4985-a236-206df00c11e1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.897052] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2b7448-a75c-4c98-a817-a5366fab60c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.939021] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7c64ed-fb0b-4ee7-9312-4a33931ffd63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.947862] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.948292] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Instance network_info: |[{"id": "bb025d74-6765-4162-98a0-e008a39320fe", "address": "fa:16:3e:a1:37:dc", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb025d74-67", "ovs_interfaceid": "bb025d74-6765-4162-98a0-e008a39320fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 983.948948] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquired lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.949080] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Refreshing network info cache for port bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.950449] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:37:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b931c4c-f73c-4fbd-9c9f-0270834cc69e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb025d74-6765-4162-98a0-e008a39320fe', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.962564] env[61947]: DEBUG oslo.service.loopingcall [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.963927] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ab0ebc-6b18-4ea6-bd04-1353a87f83b9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.975317] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 983.975317] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d23af38-d993-479c-a80d-491d667f96a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.009107] env[61947]: DEBUG nova.compute.provider_tree [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.010339] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.010339] env[61947]: value = "task-1224798" [ 984.010339] env[61947]: _type = "Task" [ 984.010339] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.021795] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224798, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.036267] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224796, 'name': ReconfigVM_Task, 'duration_secs': 0.522894} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.039311] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 85210032-a4e2-4f76-96e7-efa2fa6e134e/85210032-a4e2-4f76-96e7-efa2fa6e134e.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.040335] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-386dddbf-7d57-4d02-b93a-73d8f21c743e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.051982] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 984.051982] env[61947]: value = "task-1224799" [ 984.051982] env[61947]: _type = "Task" [ 984.051982] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.065571] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224799, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.124241] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224797, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.257855] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Updated VIF entry in instance network info cache for port bb025d74-6765-4162-98a0-e008a39320fe. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.258282] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Updating instance_info_cache with network_info: [{"id": "bb025d74-6765-4162-98a0-e008a39320fe", "address": "fa:16:3e:a1:37:dc", "network": {"id": "0815faee-1ad6-4050-8dd4-1c269fd11fa1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-116591909-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51aa57efe97e453783044286f33d3f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b931c4c-f73c-4fbd-9c9f-0270834cc69e", "external-id": "nsx-vlan-transportzone-177", "segmentation_id": 177, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb025d74-67", "ovs_interfaceid": "bb025d74-6765-4162-98a0-e008a39320fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.512032] env[61947]: DEBUG nova.scheduler.client.report [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.525042] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224798, 'name': CreateVM_Task, 'duration_secs': 0.439966} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.525042] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 984.525608] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.525778] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.526111] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.527050] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca4707ec-9009-4fc7-aba7-5bf68193725b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.534695] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 984.534695] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52529346-81a0-5481-c158-d5d5859e64e6" [ 984.534695] env[61947]: _type = "Task" [ 984.534695] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.544314] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52529346-81a0-5481-c158-d5d5859e64e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.561222] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224799, 'name': Rename_Task, 'duration_secs': 0.336094} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.561503] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 984.561751] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29e2964e-575d-491b-ad29-3a463e4d1f63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.568641] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 984.568641] env[61947]: value = "task-1224800" [ 984.568641] env[61947]: _type = "Task" [ 984.568641] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.577233] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.629012] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224797, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519784} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.629323] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ebf276db-cc9b-441f-a01b-7f7dc5b83fd3/ebf276db-cc9b-441f-a01b-7f7dc5b83fd3.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 984.629537] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.629798] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08b8293d-a2bc-426a-be93-6639276bf26f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.638038] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 984.638038] env[61947]: value = "task-1224801" [ 984.638038] env[61947]: _type = "Task" [ 984.638038] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.647532] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.761264] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Releasing lock "refresh_cache-41f62ad6-b11e-4c77-98bc-47e26234112c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.761566] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.761743] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 984.762062] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.762151] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.762276] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.019919] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.024212] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.690s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.024738] env[61947]: DEBUG nova.objects.instance [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lazy-loading 'resources' on Instance uuid 8c342d64-75f6-4b7d-829d-889eb3b6122f {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.047314] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52529346-81a0-5481-c158-d5d5859e64e6, 'name': SearchDatastore_Task, 'duration_secs': 0.010749} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.047767] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.048039] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.048278] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.048423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.048601] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.048890] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ef156fe-c8fb-47cf-857d-e5d54e12e2b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.058620] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.058797] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.059572] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f1bf2a9-d8b4-4678-9948-ef7d0f07c442 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.066471] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 985.066471] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521029f3-a746-c2f7-c9f3-0894d6048c53" [ 985.066471] env[61947]: _type = "Task" [ 985.066471] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.078167] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.081628] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521029f3-a746-c2f7-c9f3-0894d6048c53, 'name': SearchDatastore_Task, 'duration_secs': 0.009825} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.082340] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e9a4628-d90a-4073-80ce-6ee4046e3447 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.088047] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 985.088047] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254308f-892e-1bb2-f25e-88ce592c0cc8" [ 985.088047] env[61947]: _type = "Task" [ 985.088047] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.096745] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254308f-892e-1bb2-f25e-88ce592c0cc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.148718] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073007} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.149075] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.152055] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5271cda-8768-4f72-ad48-f0d347e8add6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.175996] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] ebf276db-cc9b-441f-a01b-7f7dc5b83fd3/ebf276db-cc9b-441f-a01b-7f7dc5b83fd3.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.176366] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0621026-7d30-4a61-b59b-40f233a55b86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.201433] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 985.201433] env[61947]: value = "task-1224802" [ 985.201433] env[61947]: _type = "Task" [ 985.201433] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.209846] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224802, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.544606] env[61947]: DEBUG oslo_concurrency.lockutils [None req-75fe3b5f-f374-4762-a535-9c89f261e601 tempest-ServerActionsV293TestJSON-1361578960 tempest-ServerActionsV293TestJSON-1361578960-project-member] Lock "91205e54-6bcb-403c-8308-0f0692fe4fba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.477s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.551358] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.551781] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.583326] env[61947]: DEBUG oslo_vmware.api [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224800, 'name': PowerOnVM_Task, 'duration_secs': 0.935418} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.586146] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 985.586453] env[61947]: INFO nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Took 8.66 seconds to spawn the instance on the hypervisor. [ 985.586702] env[61947]: DEBUG nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.588302] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07f7da9-ba99-44eb-bccc-e26b374ff3e7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.604856] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254308f-892e-1bb2-f25e-88ce592c0cc8, 'name': SearchDatastore_Task, 'duration_secs': 0.009826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.612011] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.612011] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 41f62ad6-b11e-4c77-98bc-47e26234112c/41f62ad6-b11e-4c77-98bc-47e26234112c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 985.612011] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fd18aa6-c4e6-4ea6-86d6-9139c612edb7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.617440] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 985.617440] env[61947]: value = "task-1224803" [ 985.617440] env[61947]: _type = "Task" [ 985.617440] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.629768] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.715629] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224802, 'name': ReconfigVM_Task, 'duration_secs': 0.363593} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.716704] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Reconfigured VM instance instance-00000058 to attach disk [datastore2] ebf276db-cc9b-441f-a01b-7f7dc5b83fd3/ebf276db-cc9b-441f-a01b-7f7dc5b83fd3.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.717691] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f78c98df-7fb9-480a-903e-987ddcefe247 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.728747] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 985.728747] env[61947]: value = "task-1224804" [ 985.728747] env[61947]: _type = "Task" [ 985.728747] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.738730] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224804, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.754019] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3321167d-2100-416e-baa1-348541dd1f00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.773191] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd101ee-4164-4917-bcb7-785df1d58e21 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.780833] env[61947]: DEBUG nova.compute.manager [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.781830] env[61947]: DEBUG nova.compute.manager [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing instance network info cache due to event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 985.781830] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.782023] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.782220] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.815556] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93be4a4-f3e0-4b4e-b0a4-0f30318f4d87 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.824763] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e8ae61-1d5d-4f47-b1ea-ebdfc811ca03 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.839494] env[61947]: DEBUG nova.compute.provider_tree [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.056552] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.056552] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.056552] env[61947]: DEBUG nova.compute.manager [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing instance network info cache due to event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 986.056552] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.119036] env[61947]: INFO nova.compute.manager [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Took 20.68 seconds to build instance. [ 986.135773] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224803, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.237514] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224804, 'name': Rename_Task, 'duration_secs': 0.165802} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.237909] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.238531] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59d24fb7-37a8-48fc-8e78-71b913ee92fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.246222] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 986.246222] env[61947]: value = "task-1224805" [ 986.246222] env[61947]: _type = "Task" [ 986.246222] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.254979] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.342467] env[61947]: DEBUG nova.scheduler.client.report [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.510535] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267e1070-10a2-47f9-9ec3-ccf3707307fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.518326] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Suspending the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 986.518458] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-85c1a245-9a8d-478a-88fc-e63ef3b02c56 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.521377] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updated VIF entry in instance network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.521743] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.525879] env[61947]: DEBUG oslo_vmware.api [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 986.525879] env[61947]: value = "task-1224806" [ 986.525879] env[61947]: _type = "Task" [ 986.525879] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.535222] env[61947]: DEBUG oslo_vmware.api [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224806, 'name': SuspendVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.622093] env[61947]: DEBUG oslo_concurrency.lockutils [None req-f953b8fa-e67e-4be9-bc75-afbc98975062 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.197s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.631913] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632814} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.632211] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 41f62ad6-b11e-4c77-98bc-47e26234112c/41f62ad6-b11e-4c77-98bc-47e26234112c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 986.632434] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.632705] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00480d42-ea7e-4a5a-bb87-210d6e6ab3de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.639843] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 986.639843] env[61947]: value = "task-1224807" [ 986.639843] env[61947]: _type = "Task" [ 986.639843] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.650065] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.758253] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224805, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.847895] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.850650] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.908s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.850910] env[61947]: DEBUG nova.objects.instance [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'resources' on Instance uuid c4647324-7e52-44d4-a4b0-961684cc1c4c {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.871833] env[61947]: INFO nova.scheduler.client.report [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Deleted allocations for instance 8c342d64-75f6-4b7d-829d-889eb3b6122f [ 987.024932] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.025332] env[61947]: DEBUG nova.compute.manager [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.025565] env[61947]: DEBUG nova.compute.manager [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 987.025856] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.026075] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.026307] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.027870] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.028129] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.042760] env[61947]: DEBUG oslo_vmware.api [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224806, 'name': SuspendVM_Task} progress is 37%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.149996] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13592} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.150306] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.151329] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850f510e-444e-4bc1-a590-71de00e2464a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.175318] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 41f62ad6-b11e-4c77-98bc-47e26234112c/41f62ad6-b11e-4c77-98bc-47e26234112c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.176052] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd612d0f-28f4-4d39-b6b4-687da7e89ff2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.197704] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 987.197704] env[61947]: value = "task-1224808" [ 987.197704] env[61947]: _type = "Task" [ 987.197704] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.206860] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.258017] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224805, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.381693] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c0cdc90-b795-4c15-9878-b3fe5aa04de1 tempest-VolumesAdminNegativeTest-999989413 tempest-VolumesAdminNegativeTest-999989413-project-member] Lock "8c342d64-75f6-4b7d-829d-889eb3b6122f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.782s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.521023] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb3cbbf-c4f6-4640-ac9e-936fcde1a5c6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.527592] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbd5e33-9be0-4e2f-b2de-53ac89446cb6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.572421] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0043642-31c4-4542-bdf5-d3d3f275330c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.576997] env[61947]: DEBUG oslo_vmware.api [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224806, 'name': SuspendVM_Task} progress is 37%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.588988] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b48b7d-c1f5-43aa-a1ab-a59b845eaa3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.606037] env[61947]: DEBUG nova.compute.provider_tree [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.710201] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224808, 'name': ReconfigVM_Task, 'duration_secs': 0.302878} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.710519] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 41f62ad6-b11e-4c77-98bc-47e26234112c/41f62ad6-b11e-4c77-98bc-47e26234112c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.711192] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cd813f8-c5fe-4573-bd50-34de8e84f365 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.720453] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 987.720453] env[61947]: value = "task-1224809" [ 987.720453] env[61947]: _type = "Task" [ 987.720453] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.729159] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224809, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.761989] env[61947]: DEBUG oslo_vmware.api [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224805, 'name': PowerOnVM_Task, 'duration_secs': 1.027351} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.762652] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.762781] env[61947]: INFO nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Took 8.58 seconds to spawn the instance on the hypervisor. [ 987.762987] env[61947]: DEBUG nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.763883] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4346c7-891a-4ee9-b109-8f5f91c546c6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.800386] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updated VIF entry in instance network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.800835] env[61947]: DEBUG nova.network.neutron [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.927018] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.927433] env[61947]: DEBUG nova.network.neutron [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.042072] env[61947]: DEBUG oslo_vmware.api [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224806, 'name': SuspendVM_Task, 'duration_secs': 1.165964} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.042513] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Suspended the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 988.042702] env[61947]: DEBUG nova.compute.manager [None req-3e141ba4-dad0-4261-8ebd-a344f0d22fab tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.043750] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8795515a-60df-40f3-b649-b5f5f636340c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.109633] env[61947]: DEBUG nova.scheduler.client.report [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.234425] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224809, 'name': Rename_Task, 'duration_secs': 0.150013} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.234700] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 988.234951] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5c4a249-c997-481e-b0a1-e15b6f12a4ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.242803] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 988.242803] env[61947]: value = "task-1224810" [ 988.242803] env[61947]: _type = "Task" [ 988.242803] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.251844] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.286968] env[61947]: INFO nova.compute.manager [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Took 20.87 seconds to build instance. [ 988.304348] env[61947]: DEBUG oslo_concurrency.lockutils [req-df51efcb-d0f6-4abb-93dd-183f8ccd9bd9 req-3193f27c-e951-4479-9d03-da30ea1095af service nova] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.433027] env[61947]: DEBUG oslo_concurrency.lockutils [req-93f7d543-1c05-4830-9c95-227dbfeb6f52 req-8c2ca830-b97b-4985-af6b-be30bf63789a service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.617408] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.621024] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.124s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.621024] env[61947]: INFO nova.compute.claims [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.657370] env[61947]: INFO nova.scheduler.client.report [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted allocations for instance c4647324-7e52-44d4-a4b0-961684cc1c4c [ 988.665222] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.755138] env[61947]: DEBUG oslo_vmware.api [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224810, 'name': PowerOnVM_Task, 'duration_secs': 0.489911} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.756049] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.762085] env[61947]: INFO nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Took 7.68 seconds to spawn the instance on the hypervisor. [ 988.762085] env[61947]: DEBUG nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.762085] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9471aa82-3f12-419a-bcd5-795d70e84984 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.790223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-63474a8a-208f-4004-bae9-69d8ad07b5a8 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.385s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.791733] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.127s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.792168] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.794017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.794017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.795215] env[61947]: INFO nova.compute.manager [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Terminating instance [ 988.797919] env[61947]: DEBUG nova.compute.manager [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.798256] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 988.800068] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303dbf04-6174-45a0-88c7-7cdfcc54021c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.811028] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 988.812549] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b973d192-92cd-46fa-8943-8cf48dcce62b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.820144] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 988.820144] env[61947]: value = "task-1224811" [ 988.820144] env[61947]: _type = "Task" [ 988.820144] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.830374] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.104804] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.105105] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.105368] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.105583] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.105742] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.108112] env[61947]: INFO nova.compute.manager [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Terminating instance [ 989.113407] env[61947]: DEBUG nova.compute.manager [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.113637] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 989.114741] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0e19af-53e8-49c8-a6aa-2345eb3ae6d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.129012] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 989.129547] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3002b123-6bd6-40e2-b635-eab33b00b3b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.168305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-95d1b70c-b5f3-47be-a967-d50e9ee8a416 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "c4647324-7e52-44d4-a4b0-961684cc1c4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.921s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.210838] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.211290] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.211408] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleting the datastore file [datastore2] 85210032-a4e2-4f76-96e7-efa2fa6e134e {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.211636] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b117320c-4160-4036-a0f5-ad3f83143ca4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.219839] env[61947]: DEBUG oslo_vmware.api [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 989.219839] env[61947]: value = "task-1224813" [ 989.219839] env[61947]: _type = "Task" [ 989.219839] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.234939] env[61947]: DEBUG oslo_vmware.api [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.286749] env[61947]: INFO nova.compute.manager [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Took 20.99 seconds to build instance. [ 989.332332] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224811, 'name': PowerOffVM_Task, 'duration_secs': 0.256455} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.332684] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 989.332804] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 989.333091] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2def8fed-c6a5-4b14-a413-3d20c4559262 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.404478] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.404740] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.404927] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Deleting the datastore file [datastore2] ebf276db-cc9b-441f-a01b-7f7dc5b83fd3 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.405304] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2107f91-8e6c-4a3d-9e35-09bb782e1c47 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.413656] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for the task: (returnval){ [ 989.413656] env[61947]: value = "task-1224815" [ 989.413656] env[61947]: _type = "Task" [ 989.413656] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.423576] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.736365] env[61947]: DEBUG oslo_vmware.api [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352996} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.738140] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.738488] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 989.739882] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 989.740229] env[61947]: INFO nova.compute.manager [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Took 0.63 seconds to destroy the instance on the hypervisor. [ 989.743616] env[61947]: DEBUG oslo.service.loopingcall [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.743616] env[61947]: DEBUG nova.compute.manager [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.743616] env[61947]: DEBUG nova.network.neutron [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.774931] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422b04ed-e0c6-4527-93e2-727f3bc287f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.782652] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee7337d-0ca7-4ec8-9742-c2be3a413595 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.788604] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b5b143a5-196d-45ec-8253-39dd06ff316f tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.497s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.818873] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bda3a5-aa04-4c61-8c56-aca0e130129a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.827212] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbf272f-4df2-484e-b45f-ae4845e0fdd8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.841443] env[61947]: DEBUG nova.compute.provider_tree [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.925705] env[61947]: DEBUG oslo_vmware.api [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Task: {'id': task-1224815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.453487} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.926064] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.926283] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 989.926709] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 989.926984] env[61947]: INFO nova.compute.manager [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 989.927271] env[61947]: DEBUG oslo.service.loopingcall [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.927582] env[61947]: DEBUG nova.compute.manager [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.927582] env[61947]: DEBUG nova.network.neutron [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 990.193759] env[61947]: DEBUG nova.compute.manager [req-8b3e2eab-96c3-4e0d-b19e-c42f9775cffc req-fc9d617f-cfab-4739-9cc9-7aa842966576 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Received event network-vif-deleted-b3e0634e-84a9-477c-b0fd-e027235d9cf4 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.194382] env[61947]: INFO nova.compute.manager [req-8b3e2eab-96c3-4e0d-b19e-c42f9775cffc req-fc9d617f-cfab-4739-9cc9-7aa842966576 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Neutron deleted interface b3e0634e-84a9-477c-b0fd-e027235d9cf4; detaching it from the instance and deleting it from the info cache [ 990.194382] env[61947]: DEBUG nova.network.neutron [req-8b3e2eab-96c3-4e0d-b19e-c42f9775cffc req-fc9d617f-cfab-4739-9cc9-7aa842966576 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.339054] env[61947]: DEBUG nova.compute.manager [req-249ff162-1ab9-4126-a7c2-cdd06926ac6a req-28652845-bcba-4b45-94c0-9b054ab5a1b6 service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Received event network-vif-deleted-76040c0b-35be-42ca-92e1-56f8ebc53495 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.339383] env[61947]: INFO nova.compute.manager [req-249ff162-1ab9-4126-a7c2-cdd06926ac6a req-28652845-bcba-4b45-94c0-9b054ab5a1b6 service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Neutron deleted interface 76040c0b-35be-42ca-92e1-56f8ebc53495; detaching it from the instance and deleting it from the info cache [ 990.339751] env[61947]: DEBUG nova.network.neutron [req-249ff162-1ab9-4126-a7c2-cdd06926ac6a req-28652845-bcba-4b45-94c0-9b054ab5a1b6 service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.347517] env[61947]: DEBUG nova.scheduler.client.report [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.658262] env[61947]: DEBUG nova.network.neutron [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.696850] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32b84fb5-9a2f-456c-88f3-4f7f1839004c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.707633] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391cb5e0-083b-47f5-8e5b-d84722cd56f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.735476] env[61947]: DEBUG nova.compute.manager [req-8b3e2eab-96c3-4e0d-b19e-c42f9775cffc req-fc9d617f-cfab-4739-9cc9-7aa842966576 service nova] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Detach interface failed, port_id=b3e0634e-84a9-477c-b0fd-e027235d9cf4, reason: Instance ebf276db-cc9b-441f-a01b-7f7dc5b83fd3 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 990.816679] env[61947]: DEBUG nova.network.neutron [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.855013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.855013] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.856350] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e717e0d-8adf-4a34-8ae1-74a631e6fab8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.867758] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b85e67-0677-4bc1-ac68-7d5a82f453f8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.903523] env[61947]: DEBUG nova.compute.manager [req-249ff162-1ab9-4126-a7c2-cdd06926ac6a req-28652845-bcba-4b45-94c0-9b054ab5a1b6 service nova] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Detach interface failed, port_id=76040c0b-35be-42ca-92e1-56f8ebc53495, reason: Instance 85210032-a4e2-4f76-96e7-efa2fa6e134e could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 990.961571] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "41f62ad6-b11e-4c77-98bc-47e26234112c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.961820] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.961992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.962319] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.962375] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.965025] env[61947]: INFO nova.compute.manager [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Terminating instance [ 990.967952] env[61947]: DEBUG nova.compute.manager [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.967952] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.968467] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7623152-915d-454d-892b-4fadaa60326c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.977799] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.978073] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26585d50-9fa8-40a8-9794-6ca4bbb62c3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.985694] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 990.985694] env[61947]: value = "task-1224816" [ 990.985694] env[61947]: _type = "Task" [ 990.985694] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.995781] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.157254] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.157604] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.159908] env[61947]: INFO nova.compute.manager [-] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Took 1.23 seconds to deallocate network for instance. [ 991.318333] env[61947]: INFO nova.compute.manager [-] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Took 1.58 seconds to deallocate network for instance. [ 991.361676] env[61947]: DEBUG nova.compute.utils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.363046] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.363159] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.423030] env[61947]: DEBUG nova.policy [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a0420988621c4dd39cfbc99db4b9719e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d76ebc8875f4b03bc768320b44b11d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 991.496108] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224816, 'name': PowerOffVM_Task, 'duration_secs': 0.17014} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.496402] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.496577] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.496838] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b016cf65-879d-4dc4-944c-0ae639b74aff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.565077] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.565385] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.565677] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleting the datastore file [datastore2] 41f62ad6-b11e-4c77-98bc-47e26234112c {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.565886] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-351279b1-59a9-44a9-bc68-1bc5d6cb4d3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.575071] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for the task: (returnval){ [ 991.575071] env[61947]: value = "task-1224818" [ 991.575071] env[61947]: _type = "Task" [ 991.575071] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.583765] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.665424] env[61947]: INFO nova.compute.manager [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Detaching volume 51d58e15-9bd1-43c0-a542-8210c2873e98 [ 991.668795] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.669223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.670047] env[61947]: DEBUG nova.objects.instance [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lazy-loading 'resources' on Instance uuid ebf276db-cc9b-441f-a01b-7f7dc5b83fd3 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.728623] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Successfully created port: 487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.732626] env[61947]: INFO nova.virt.block_device [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Attempting to driver detach volume 51d58e15-9bd1-43c0-a542-8210c2873e98 from mountpoint /dev/sdb [ 991.733644] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 991.733644] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264684', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'name': 'volume-51d58e15-9bd1-43c0-a542-8210c2873e98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd84b03e-ef6d-41e0-b48c-b8e8fc425a36', 'attached_at': '', 'detached_at': '', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'serial': '51d58e15-9bd1-43c0-a542-8210c2873e98'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 991.736345] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a800924-e5cd-444b-8448-b785236fbb5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.762274] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddec506-ba91-4063-a75f-a22db313cbd0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.771037] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcc6f94-87cd-4726-b4b6-4705eba0fb00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.799227] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6242d3-602f-4030-bc59-8176af25711b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.819852] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] The volume has not been displaced from its original location: [datastore1] volume-51d58e15-9bd1-43c0-a542-8210c2873e98/volume-51d58e15-9bd1-43c0-a542-8210c2873e98.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 991.822969] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 991.823364] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5eff0258-d01f-4319-9ad6-b36264433e22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.837971] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.845881] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 991.845881] env[61947]: value = "task-1224819" [ 991.845881] env[61947]: _type = "Task" [ 991.845881] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.855421] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224819, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.867236] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.086080] env[61947]: DEBUG oslo_vmware.api [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Task: {'id': task-1224818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175594} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.086361] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.086554] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.086737] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.086942] env[61947]: INFO nova.compute.manager [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 992.087226] env[61947]: DEBUG oslo.service.loopingcall [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.087435] env[61947]: DEBUG nova.compute.manager [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.087571] env[61947]: DEBUG nova.network.neutron [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.328135] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e64dac-63e1-49d3-b9e5-a44e3321b3c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.337730] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13b037c-13d8-4f8c-ae40-354f62790408 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.374939] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f15ec1-ce62-4732-895e-3b4d8054cb44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.390301] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224819, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.393590] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0de235-610a-46fb-84ff-530ee08e9ffe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.408851] env[61947]: DEBUG nova.compute.provider_tree [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.435543] env[61947]: DEBUG nova.compute.manager [req-4ff3a695-9d21-48ae-adeb-3b68819f3aa3 req-cb9c9ffb-a403-430c-9644-f66f0764eb5f service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Received event network-vif-deleted-bb025d74-6765-4162-98a0-e008a39320fe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.435756] env[61947]: INFO nova.compute.manager [req-4ff3a695-9d21-48ae-adeb-3b68819f3aa3 req-cb9c9ffb-a403-430c-9644-f66f0764eb5f service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Neutron deleted interface bb025d74-6765-4162-98a0-e008a39320fe; detaching it from the instance and deleting it from the info cache [ 992.436025] env[61947]: DEBUG nova.network.neutron [req-4ff3a695-9d21-48ae-adeb-3b68819f3aa3 req-cb9c9ffb-a403-430c-9644-f66f0764eb5f service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.881721] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224819, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.887975] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.910148] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.910437] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.910601] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.910787] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.912137] env[61947]: DEBUG nova.virt.hardware [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.912868] env[61947]: DEBUG nova.scheduler.client.report [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.915822] env[61947]: DEBUG nova.network.neutron [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.917331] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c3d1f9-9f85-44b8-8acc-2bbcedf1cce5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.927022] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743ad957-88a1-41e6-8660-49237bc84aa7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.944225] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-344f6a2c-8d9f-4cc3-84fd-8518dc4a4cee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.953797] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532f8348-68e9-426d-b6ce-6e83d1612a25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.982731] env[61947]: DEBUG nova.compute.manager [req-4ff3a695-9d21-48ae-adeb-3b68819f3aa3 req-cb9c9ffb-a403-430c-9644-f66f0764eb5f service nova] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Detach interface failed, port_id=bb025d74-6765-4162-98a0-e008a39320fe, reason: Instance 41f62ad6-b11e-4c77-98bc-47e26234112c could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 993.211773] env[61947]: DEBUG nova.compute.manager [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.212111] env[61947]: DEBUG oslo_concurrency.lockutils [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.212415] env[61947]: DEBUG oslo_concurrency.lockutils [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.212585] env[61947]: DEBUG oslo_concurrency.lockutils [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.212774] env[61947]: DEBUG nova.compute.manager [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] No waiting events found dispatching network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.212954] env[61947]: WARNING nova.compute.manager [req-4a080566-e964-48f8-ad9d-aa4d44117801 req-5c9c2489-8dee-46cc-9cd9-aa394b4690cc service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received unexpected event network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 for instance with vm_state building and task_state spawning. [ 993.354450] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Successfully updated port: 487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.382700] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224819, 'name': ReconfigVM_Task, 'duration_secs': 1.361621} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.383320] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 993.390048] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa107520-a234-4137-9001-7fc75f42f351 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.406895] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 993.406895] env[61947]: value = "task-1224820" [ 993.406895] env[61947]: _type = "Task" [ 993.406895] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.416088] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.421581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.423965] env[61947]: INFO nova.compute.manager [-] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Took 1.34 seconds to deallocate network for instance. [ 993.424530] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.587s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.425402] env[61947]: DEBUG nova.objects.instance [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'resources' on Instance uuid 85210032-a4e2-4f76-96e7-efa2fa6e134e {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.454502] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "1660df2a-b2c5-469f-b602-0014dd3656d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.454618] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.457192] env[61947]: INFO nova.scheduler.client.report [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Deleted allocations for instance ebf276db-cc9b-441f-a01b-7f7dc5b83fd3 [ 993.704932] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "8a4b81d3-d683-4517-806c-88b41ad73c48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.705236] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.810847] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.811160] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.811542] env[61947]: DEBUG nova.objects.instance [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'flavor' on Instance uuid fcd2e134-4d5f-43d5-a09c-55a49f8ce48b {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.857088] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.858217] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.858217] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.925025] env[61947]: DEBUG oslo_vmware.api [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224820, 'name': ReconfigVM_Task, 'duration_secs': 0.150887} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.925025] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264684', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'name': 'volume-51d58e15-9bd1-43c0-a542-8210c2873e98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd84b03e-ef6d-41e0-b48c-b8e8fc425a36', 'attached_at': '', 'detached_at': '', 'volume_id': '51d58e15-9bd1-43c0-a542-8210c2873e98', 'serial': '51d58e15-9bd1-43c0-a542-8210c2873e98'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 993.937614] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.956706] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 993.968217] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c32b4928-466d-4eb0-8b05-7835b507aa08 tempest-ImagesNegativeTestJSON-1107464169 tempest-ImagesNegativeTestJSON-1107464169-project-member] Lock "ebf276db-cc9b-441f-a01b-7f7dc5b83fd3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.173s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.112234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "4e0a8c79-9445-4474-b516-e697e1854dea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.112234] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.128613] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a219bffd-e707-4ad8-906e-980aef23b193 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.138234] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab82565-704b-4f21-966a-402640833c83 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.174220] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f395a7cc-103c-4249-8a9a-bbb5f8e747d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.183916] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532fbe55-1b2a-4801-947f-46a6afafa65e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.204809] env[61947]: DEBUG nova.compute.provider_tree [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.207274] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 994.408033] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.428242] env[61947]: DEBUG nova.objects.instance [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'pci_requests' on Instance uuid fcd2e134-4d5f-43d5-a09c-55a49f8ce48b {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.480339] env[61947]: DEBUG nova.objects.instance [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.488052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.619222] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 994.639233] env[61947]: DEBUG nova.network.neutron [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.706902] env[61947]: DEBUG nova.scheduler.client.report [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.732195] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.931436] env[61947]: DEBUG nova.objects.base [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 994.931750] env[61947]: DEBUG nova.network.neutron [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 995.076781] env[61947]: DEBUG nova.policy [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 995.142642] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.143233] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.143527] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance network_info: |[{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.143931] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:65:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '487b8623-6225-453d-9be0-1dd6e1241b97', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.151414] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating folder: Project (1d76ebc8875f4b03bc768320b44b11d1). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 995.151735] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4456e0a9-de93-44ce-8e3e-ebcdd0c4805c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.165117] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created folder: Project (1d76ebc8875f4b03bc768320b44b11d1) in parent group-v264556. [ 995.165368] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating folder: Instances. Parent ref: group-v264714. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 995.165822] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94d49525-131c-4c40-b3fd-20c41bfd5838 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.177450] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created folder: Instances in parent group-v264714. [ 995.177719] env[61947]: DEBUG oslo.service.loopingcall [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.177921] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.178156] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1924c8fc-19f6-4887-98cc-c7929a3ca53a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.199805] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.199805] env[61947]: value = "task-1224823" [ 995.199805] env[61947]: _type = "Task" [ 995.199805] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.210094] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224823, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.217173] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.220017] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.283s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.220299] env[61947]: DEBUG nova.objects.instance [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lazy-loading 'resources' on Instance uuid 41f62ad6-b11e-4c77-98bc-47e26234112c {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.249724] env[61947]: INFO nova.scheduler.client.report [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance 85210032-a4e2-4f76-96e7-efa2fa6e134e [ 995.391027] env[61947]: DEBUG nova.compute.manager [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.391348] env[61947]: DEBUG nova.compute.manager [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing instance network info cache due to event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 995.391691] env[61947]: DEBUG oslo_concurrency.lockutils [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.391885] env[61947]: DEBUG oslo_concurrency.lockutils [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.392372] env[61947]: DEBUG nova.network.neutron [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.488092] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fcf1128c-f2fe-4506-97eb-23bd28ec64f4 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.330s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.714653] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224823, 'name': CreateVM_Task, 'duration_secs': 0.332636} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.714653] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.714653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.714653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.714653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.714653] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22124365-f090-4665-971d-03281e472b85 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.721324] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 995.721324] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52150f95-a02e-a04b-f23d-75ae37f4b826" [ 995.721324] env[61947]: _type = "Task" [ 995.721324] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.731729] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52150f95-a02e-a04b-f23d-75ae37f4b826, 'name': SearchDatastore_Task, 'duration_secs': 0.010329} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.732189] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.732572] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.733042] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.733369] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.733725] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.734154] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d727bd42-6768-4552-af93-34a6a7ed63ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.745023] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.745023] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.745023] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04792cf9-7cc6-4641-9809-daea80587f11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.753449] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 995.753449] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267d77d-b1b5-2d7d-3f7e-6ae836453c83" [ 995.753449] env[61947]: _type = "Task" [ 995.753449] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.763890] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267d77d-b1b5-2d7d-3f7e-6ae836453c83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.764846] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f769773-38fe-4251-80c6-a98269970693 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "85210032-a4e2-4f76-96e7-efa2fa6e134e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.659s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.881879] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad2a364-3d36-4374-bbbe-cbccb2856640 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.891127] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c58ae75-1041-4163-ac65-48b11f763dbf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.930106] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0fc388-9d2f-40a7-b69e-afd74762a6fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.940960] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ec8c6c-9c52-48bc-b15a-c38c82e9a007 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.957056] env[61947]: DEBUG nova.compute.provider_tree [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.115931] env[61947]: DEBUG nova.network.neutron [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updated VIF entry in instance network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.116447] env[61947]: DEBUG nova.network.neutron [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.261373] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5267d77d-b1b5-2d7d-3f7e-6ae836453c83, 'name': SearchDatastore_Task, 'duration_secs': 0.021694} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.262172] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1590e5c-809e-4988-ac50-c3d807aeac6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.267716] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 996.267716] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520d11e4-71b4-02c7-59b6-759b07ad7181" [ 996.267716] env[61947]: _type = "Task" [ 996.267716] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.276064] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520d11e4-71b4-02c7-59b6-759b07ad7181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.421998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.421998] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.460432] env[61947]: DEBUG nova.scheduler.client.report [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.529407] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.529689] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.529933] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.530183] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.530387] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.532515] env[61947]: INFO nova.compute.manager [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Terminating instance [ 996.534275] env[61947]: DEBUG nova.compute.manager [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.534473] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 996.535394] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cd2743-26db-437b-ab90-4e9e3165baea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.544666] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 996.547944] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06b88f9b-a202-41fd-9210-fb1e93c76e3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.553438] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 996.553438] env[61947]: value = "task-1224824" [ 996.553438] env[61947]: _type = "Task" [ 996.553438] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.566601] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.619343] env[61947]: DEBUG oslo_concurrency.lockutils [req-f8e4b975-228b-4d8f-a721-68367b5f4700 req-a005911e-87a5-4c75-918e-e2cc0cb79b5b service nova] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.632984] env[61947]: DEBUG nova.compute.manager [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.633178] env[61947]: DEBUG oslo_concurrency.lockutils [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.633400] env[61947]: DEBUG oslo_concurrency.lockutils [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.633577] env[61947]: DEBUG oslo_concurrency.lockutils [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.633744] env[61947]: DEBUG nova.compute.manager [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] No waiting events found dispatching network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 996.633969] env[61947]: WARNING nova.compute.manager [req-4022727b-d412-4061-b0a1-8b1580f18bf5 req-0d35a70d-4558-4abb-8bf7-01d9cdd88a14 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received unexpected event network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 for instance with vm_state active and task_state None. [ 996.778472] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520d11e4-71b4-02c7-59b6-759b07ad7181, 'name': SearchDatastore_Task, 'duration_secs': 0.009859} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.778765] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.779054] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.779318] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1be16fb3-d565-47d3-b4f4-a7931bcaa023 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.786255] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 996.786255] env[61947]: value = "task-1224825" [ 996.786255] env[61947]: _type = "Task" [ 996.786255] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.799930] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.808592] env[61947]: DEBUG nova.network.neutron [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Successfully updated port: 76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.926965] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.965544] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.968013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.480s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.969538] env[61947]: INFO nova.compute.claims [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.989112] env[61947]: INFO nova.scheduler.client.report [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Deleted allocations for instance 41f62ad6-b11e-4c77-98bc-47e26234112c [ 997.069167] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224824, 'name': PowerOffVM_Task, 'duration_secs': 0.197483} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.069409] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 997.069606] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 997.069904] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4db0f99f-1a1b-4e52-8f3b-487c204fafec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.139085] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 997.139671] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 997.139671] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleting the datastore file [datastore2] bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.139999] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dab40962-2bae-4999-8a17-0d189667461f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.148945] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 997.148945] env[61947]: value = "task-1224827" [ 997.148945] env[61947]: _type = "Task" [ 997.148945] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.163712] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.304172] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224825, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.313399] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.313751] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.314083] env[61947]: DEBUG nova.network.neutron [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.457337] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.499542] env[61947]: DEBUG oslo_concurrency.lockutils [None req-250c592b-c950-4e34-96bd-a53eb5166784 tempest-ServerDiskConfigTestJSON-1795548106 tempest-ServerDiskConfigTestJSON-1795548106-project-member] Lock "41f62ad6-b11e-4c77-98bc-47e26234112c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.536s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.662504] env[61947]: DEBUG oslo_vmware.api [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255719} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.662844] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.663054] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 997.663465] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 997.663719] env[61947]: INFO nova.compute.manager [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Took 1.13 seconds to destroy the instance on the hypervisor. [ 997.663977] env[61947]: DEBUG oslo.service.loopingcall [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.664197] env[61947]: DEBUG nova.compute.manager [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.664288] env[61947]: DEBUG nova.network.neutron [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.799510] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561574} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.799792] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.800030] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.800308] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b62485e-0322-4c7b-8637-9ba80f8fcad5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.808601] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 997.808601] env[61947]: value = "task-1224828" [ 997.808601] env[61947]: _type = "Task" [ 997.808601] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.820748] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224828, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.893789] env[61947]: WARNING nova.network.neutron [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] 720a7469-85cf-4e25-82f3-b574730aae1f already exists in list: networks containing: ['720a7469-85cf-4e25-82f3-b574730aae1f']. ignoring it [ 998.132945] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f89c04-6fd1-44cd-b819-a759ef50d000 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.142800] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a30f79-3f3f-4606-9b3b-2e15f554537f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.177740] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5816671-a761-4892-b250-7a2473c3a05a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.187137] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b307f270-d68c-485b-b55b-ae89b6d79e91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.203936] env[61947]: DEBUG nova.compute.provider_tree [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.323555] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224828, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074346} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.323890] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.327924] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d599e954-3e91-4a81-8803-53099beebc0b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.116545] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.121515] env[61947]: DEBUG nova.compute.manager [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.121703] env[61947]: DEBUG nova.compute.manager [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-76c0251c-9621-4df0-83f5-b4a19d1f5da9. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 999.121896] env[61947]: DEBUG oslo_concurrency.lockutils [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.122416] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3907be24-beb5-4876-9125-e0d891773d31 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.137319] env[61947]: ERROR nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [req-ee014287-143d-43a0-a41e-688c74525bc3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ee014287-143d-43a0-a41e-688c74525bc3"}]} [ 999.138304] env[61947]: DEBUG nova.network.neutron [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.149071] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 999.149071] env[61947]: value = "task-1224829" [ 999.149071] env[61947]: _type = "Task" [ 999.149071] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.158952] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224829, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.160768] env[61947]: DEBUG nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 999.187716] env[61947]: DEBUG nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 999.187943] env[61947]: DEBUG nova.compute.provider_tree [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.213440] env[61947]: DEBUG nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 999.237689] env[61947]: DEBUG nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 999.426696] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdd2de1-6075-4262-b73d-823fa3d80a5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.431945] env[61947]: DEBUG nova.network.neutron [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "address": "fa:16:3e:85:91:f7", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76c0251c-96", "ovs_interfaceid": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.436179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d9950d-02e1-421c-ab87-a96b84736d73 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.472014] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db410b5-4dd7-419c-9cb3-355aae948bf0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.483995] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7733d6ad-5f9c-42e2-81ab-319f247e5317 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.496905] env[61947]: DEBUG nova.compute.provider_tree [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.642711] env[61947]: INFO nova.compute.manager [-] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Took 1.98 seconds to deallocate network for instance. [ 999.663083] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224829, 'name': ReconfigVM_Task, 'duration_secs': 0.269826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.663083] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.663083] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed3aa385-ba93-406d-b07b-a4c444fab75d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.669081] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 999.669081] env[61947]: value = "task-1224830" [ 999.669081] env[61947]: _type = "Task" [ 999.669081] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.678194] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224830, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.939036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.939036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.939036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.939036] env[61947]: DEBUG oslo_concurrency.lockutils [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.939036] env[61947]: DEBUG nova.network.neutron [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port 76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.940578] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c130bb-f612-44fe-8cc3-0a5f6368936d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.966791] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.967828] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.968201] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.968710] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.969333] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.970021] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.970231] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.971073] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.971073] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.971073] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.971890] env[61947]: DEBUG nova.virt.hardware [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.979110] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfiguring VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 999.979994] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a55e57e-0e7a-4a1e-80ea-7caa55518610 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.003515] env[61947]: DEBUG oslo_vmware.api [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1000.003515] env[61947]: value = "task-1224831" [ 1000.003515] env[61947]: _type = "Task" [ 1000.003515] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.013528] env[61947]: DEBUG oslo_vmware.api [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224831, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.032884] env[61947]: DEBUG nova.scheduler.client.report [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 122 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1000.033180] env[61947]: DEBUG nova.compute.provider_tree [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 122 to 123 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1000.033370] env[61947]: DEBUG nova.compute.provider_tree [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.149596] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.183696] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224830, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.518250] env[61947]: DEBUG oslo_vmware.api [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224831, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.542023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.570s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.542023] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1000.542757] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.811s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.546804] env[61947]: INFO nova.compute.claims [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.687972] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224830, 'name': Rename_Task, 'duration_secs': 0.825871} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.690932] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.691572] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bcd2a86-57f3-4361-adcf-bc26df764df6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.698773] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1000.698773] env[61947]: value = "task-1224832" [ 1000.698773] env[61947]: _type = "Task" [ 1000.698773] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.708276] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.734654] env[61947]: DEBUG nova.network.neutron [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port 76c0251c-9621-4df0-83f5-b4a19d1f5da9. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.735133] env[61947]: DEBUG nova.network.neutron [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "address": "fa:16:3e:85:91:f7", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76c0251c-96", "ovs_interfaceid": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.845648] env[61947]: DEBUG nova.compute.manager [req-06e836e6-b10c-404a-afee-8ef87f8ad45c req-c86da620-3160-4de9-9aff-12f1a6719a20 service nova] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Received event network-vif-deleted-4024dcff-f052-4a2e-9fab-c3d722069bbe {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.019065] env[61947]: DEBUG oslo_vmware.api [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224831, 'name': ReconfigVM_Task, 'duration_secs': 0.694125} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.019693] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.019983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfigured VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1001.058183] env[61947]: DEBUG nova.compute.utils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.062999] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.063191] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1001.120230] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.120822] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.134228] env[61947]: DEBUG nova.policy [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53ae249274c44538889225db2bced1e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb9a821c72e147f7ba7e427113148af6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1001.214194] env[61947]: DEBUG oslo_vmware.api [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224832, 'name': PowerOnVM_Task, 'duration_secs': 0.456601} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.214508] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.214720] env[61947]: INFO nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Took 8.33 seconds to spawn the instance on the hypervisor. [ 1001.214904] env[61947]: DEBUG nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.215821] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b04f09-87ff-44ff-a240-fa55749c610a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.241602] env[61947]: DEBUG oslo_concurrency.lockutils [req-11b7839f-d7a5-4ef4-ab02-175c6a1bb47a req-3942433c-37bf-485e-bbdb-ab47dd8762ca service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.389675] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Successfully created port: 0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1001.524719] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a79df849-be61-4a68-9111-73bd5ed11025 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.713s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.563578] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1001.623383] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.736860] env[61947]: INFO nova.compute.manager [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Took 21.27 seconds to build instance. [ 1001.746096] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a8e49c-e78f-4c99-aa5b-659f6ae4e897 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.758687] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045e90e4-9c40-450f-8da8-8cf12ae842d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.792152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9095eeef-7063-4a7a-a9f8-0db796974867 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.800748] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a04a1d-9b9f-4870-a56a-c020cac9d8bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.815500] env[61947]: DEBUG nova.compute.provider_tree [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.148168] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.238625] env[61947]: DEBUG oslo_concurrency.lockutils [None req-758bbfd4-a050-4e90-83a5-9a926b3dddff tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.778s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.322444] env[61947]: DEBUG nova.scheduler.client.report [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.408255] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.408558] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.577020] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1002.611374] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.611630] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.611787] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.611971] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.612974] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.613191] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.613416] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.613582] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.613750] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.613917] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.614106] env[61947]: DEBUG nova.virt.hardware [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.615305] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fce3960-a6a3-4861-8641-e75d4d142b5f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.627313] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b964e6e-d9f2-4616-ad07-38a9e0044995 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.829735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.829811] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1002.832751] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.690s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.834456] env[61947]: INFO nova.compute.claims [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.914394] env[61947]: INFO nova.compute.manager [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Detaching volume e87cfa37-9ff6-496d-8dc7-6d5aadba8c93 [ 1002.972384] env[61947]: DEBUG nova.compute.manager [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Received event network-vif-plugged-0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.972713] env[61947]: DEBUG oslo_concurrency.lockutils [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] Acquiring lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.972853] env[61947]: DEBUG oslo_concurrency.lockutils [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.973018] env[61947]: DEBUG oslo_concurrency.lockutils [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.973203] env[61947]: DEBUG nova.compute.manager [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] No waiting events found dispatching network-vif-plugged-0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.973378] env[61947]: WARNING nova.compute.manager [req-b82cde12-a0e8-4cb8-acad-18dad8685f9d req-eae3f18e-6f04-4c7e-aca2-935d3c117e85 service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Received unexpected event network-vif-plugged-0c5dc16b-3789-41c8-877b-380b3cde8c15 for instance with vm_state building and task_state spawning. [ 1002.974904] env[61947]: INFO nova.virt.block_device [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Attempting to driver detach volume e87cfa37-9ff6-496d-8dc7-6d5aadba8c93 from mountpoint /dev/sdb [ 1002.975143] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1002.975346] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264706', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'name': 'volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21bb0270-bc20-4ec1-9599-d676845b0dc7', 'attached_at': '', 'detached_at': '', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'serial': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1002.976638] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf639de-da7b-4f4d-9282-ae5a3c0029a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.004570] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Successfully updated port: 0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1003.006918] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460ca226-bc37-4cb8-89f9-1acbf61279ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.016340] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0c367c-2886-4781-9817-90b18a088a76 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.042634] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c07bbf4-5355-4b64-a127-02c9190fa570 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.057292] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] The volume has not been displaced from its original location: [datastore2] volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93/volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1003.062659] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.063033] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2f185c7-d029-456b-a472-47154d661ca2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.082976] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1003.082976] env[61947]: value = "task-1224833" [ 1003.082976] env[61947]: _type = "Task" [ 1003.082976] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.092246] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224833, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.303052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.303052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.338222] env[61947]: DEBUG nova.compute.utils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.342761] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1003.343009] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1003.395493] env[61947]: DEBUG nova.policy [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53ae249274c44538889225db2bced1e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb9a821c72e147f7ba7e427113148af6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1003.409182] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.409454] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.457029] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.457263] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.511677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.512178] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.512178] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.594936] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224833, 'name': ReconfigVM_Task, 'duration_secs': 0.235342} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.595242] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1003.600235] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75dd5606-e4ab-47a1-8467-cc6e0e268120 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.624547] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1003.624547] env[61947]: value = "task-1224834" [ 1003.624547] env[61947]: _type = "Task" [ 1003.624547] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.635986] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.664896] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Successfully created port: a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1003.806756] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.841813] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1003.912025] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.960028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.960174] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.961196] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb14faba-ef77-4b35-8262-d3b47f787926 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.992767] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dffadcb-2de5-40db-9942-84bf254f9831 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.027810] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfiguring VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1004.030778] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81da8ac6-05c0-4d4d-a559-355711e2fdbf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.051941] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1004.051941] env[61947]: value = "task-1224835" [ 1004.051941] env[61947]: _type = "Task" [ 1004.051941] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.064194] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.081467] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1004.118927] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926761db-9d26-4367-b101-8b111c0ddacc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.130632] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9ad3f7-412d-4ed0-8025-b8be3cd07a9f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.136528] env[61947]: DEBUG oslo_vmware.api [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224834, 'name': ReconfigVM_Task, 'duration_secs': 0.165952} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.137141] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264706', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'name': 'volume-e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21bb0270-bc20-4ec1-9599-d676845b0dc7', 'attached_at': '', 'detached_at': '', 'volume_id': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93', 'serial': 'e87cfa37-9ff6-496d-8dc7-6d5aadba8c93'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1004.171300] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe96344f-7c14-4a99-a8df-6bed1e21dd95 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.181689] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27b4cb5-726b-45d9-b031-8e2e0e90a44d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.199790] env[61947]: DEBUG nova.compute.provider_tree [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.251507] env[61947]: DEBUG nova.network.neutron [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Updating instance_info_cache with network_info: [{"id": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "address": "fa:16:3e:40:2c:52", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5dc16b-37", "ovs_interfaceid": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.335352] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.432252] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.562604] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.693626] env[61947]: DEBUG nova.objects.instance [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.703097] env[61947]: DEBUG nova.scheduler.client.report [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.753582] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.753915] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Instance network_info: |[{"id": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "address": "fa:16:3e:40:2c:52", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5dc16b-37", "ovs_interfaceid": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.754352] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:2c:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '438671d0-9468-4e44-84c1-4c0ebaa743e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c5dc16b-3789-41c8-877b-380b3cde8c15', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.762034] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Creating folder: Project (eb9a821c72e147f7ba7e427113148af6). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1004.762336] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d42c34be-5d21-4013-9877-d3b7b715153a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.774237] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Created folder: Project (eb9a821c72e147f7ba7e427113148af6) in parent group-v264556. [ 1004.774457] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Creating folder: Instances. Parent ref: group-v264717. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1004.774765] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79cfd897-0c2f-4936-9023-12ff39c87ceb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.787444] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Created folder: Instances in parent group-v264717. [ 1004.787698] env[61947]: DEBUG oslo.service.loopingcall [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.787881] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1004.788113] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31c5dc74-c56d-421f-93bc-62ca7f3ec2bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.812771] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.812771] env[61947]: value = "task-1224838" [ 1004.812771] env[61947]: _type = "Task" [ 1004.812771] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.826825] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224838, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.853026] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1004.878916] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.879227] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.879432] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.879648] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.879801] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.879952] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.880303] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.880407] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.880569] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.880738] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.880911] env[61947]: DEBUG nova.virt.hardware [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.881856] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db078eb-2332-4242-bce9-d77a937707b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.890812] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9276ef-7f93-4e13-8144-aa5e8b9b336f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.024789] env[61947]: DEBUG nova.compute.manager [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Received event network-changed-0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.025387] env[61947]: DEBUG nova.compute.manager [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Refreshing instance network info cache due to event network-changed-0c5dc16b-3789-41c8-877b-380b3cde8c15. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1005.025387] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] Acquiring lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.025387] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] Acquired lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.025625] env[61947]: DEBUG nova.network.neutron [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Refreshing network info cache for port 0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.064048] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.208199] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.208870] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.214020] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.755s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.214020] env[61947]: INFO nova.compute.claims [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.312706] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Successfully updated port: a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1005.323682] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224838, 'name': CreateVM_Task, 'duration_secs': 0.438613} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.324366] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1005.325056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.325232] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.325560] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.326054] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-565833d4-88b6-4167-abf6-52aa0a6263f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.331008] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1005.331008] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52067cc5-3eb2-9186-b70a-a1662f476a95" [ 1005.331008] env[61947]: _type = "Task" [ 1005.331008] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.340216] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52067cc5-3eb2-9186-b70a-a1662f476a95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.564880] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.627633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.707268] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3a18cf01-429c-4d63-8968-a8228fd76402 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.298s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.708618] env[61947]: DEBUG oslo_concurrency.lockutils [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.081s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.708896] env[61947]: DEBUG nova.compute.manager [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.710143] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccc897c-b5e5-41e2-8d47-c80b58366d8b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.719098] env[61947]: DEBUG nova.compute.utils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.722276] env[61947]: DEBUG nova.compute.manager [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1005.722827] env[61947]: DEBUG nova.objects.instance [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.724118] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.724279] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.728722] env[61947]: DEBUG nova.network.neutron [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Updated VIF entry in instance network info cache for port 0c5dc16b-3789-41c8-877b-380b3cde8c15. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.729043] env[61947]: DEBUG nova.network.neutron [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Updating instance_info_cache with network_info: [{"id": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "address": "fa:16:3e:40:2c:52", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5dc16b-37", "ovs_interfaceid": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.766750] env[61947]: DEBUG nova.policy [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53ae249274c44538889225db2bced1e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb9a821c72e147f7ba7e427113148af6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1005.819942] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.820501] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.820501] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.849072] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52067cc5-3eb2-9186-b70a-a1662f476a95, 'name': SearchDatastore_Task, 'duration_secs': 0.010832} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.849482] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.849785] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1005.850085] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.850353] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.850679] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1005.851098] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0904fe4-8da9-4e0b-ba8e-8aff6e286104 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.862022] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1005.863023] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1005.863463] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f6e24a8-8107-4524-be6d-9ffa3f9fa770 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.873401] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1005.873401] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd6247-43c1-3eaa-ed14-866f03d48e23" [ 1005.873401] env[61947]: _type = "Task" [ 1005.873401] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.878430] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd6247-43c1-3eaa-ed14-866f03d48e23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.065078] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.065994] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Successfully created port: 3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.225030] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.238659] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.241488] env[61947]: DEBUG oslo_concurrency.lockutils [req-eb13e436-9a9b-4848-a313-47edf8df2c02 req-bf0ce553-c686-427b-80ab-d8c258306bad service nova] Releasing lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.241488] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a255502-ca8f-4b7b-9ca5-39ba088c14be {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.249585] env[61947]: DEBUG oslo_vmware.api [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1006.249585] env[61947]: value = "task-1224839" [ 1006.249585] env[61947]: _type = "Task" [ 1006.249585] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.267903] env[61947]: DEBUG oslo_vmware.api [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.364958] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1006.380447] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cd6247-43c1-3eaa-ed14-866f03d48e23, 'name': SearchDatastore_Task, 'duration_secs': 0.019049} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.383628] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89aff546-4ca7-40cf-be3f-7a3cb7dce819 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.391597] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1006.391597] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f9fde6-335a-8a39-055b-840f154adb2f" [ 1006.391597] env[61947]: _type = "Task" [ 1006.391597] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.403672] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f9fde6-335a-8a39-055b-840f154adb2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.422210] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7341d949-4100-4b6d-b84f-9bcae4fc8580 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.433815] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a926ed-9dc7-47f5-b2ab-cead56b197d1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.466821] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aaff8a0-4ce5-4d80-a68c-e28724f430b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.477443] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28382296-debb-4e9c-ba2f-28045300f2ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.493369] env[61947]: DEBUG nova.compute.provider_tree [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.565417] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.577129] env[61947]: DEBUG nova.network.neutron [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Updating instance_info_cache with network_info: [{"id": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "address": "fa:16:3e:b1:6d:dd", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b4c8f3-14", "ovs_interfaceid": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.762243] env[61947]: DEBUG oslo_vmware.api [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224839, 'name': PowerOffVM_Task, 'duration_secs': 0.28126} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.762520] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1006.762697] env[61947]: DEBUG nova.compute.manager [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.764238] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b09087-78a8-418a-9cb9-dbded6a64adb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.902474] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f9fde6-335a-8a39-055b-840f154adb2f, 'name': SearchDatastore_Task, 'duration_secs': 0.015405} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.902756] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.903086] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 1660df2a-b2c5-469f-b602-0014dd3656d5/1660df2a-b2c5-469f-b602-0014dd3656d5.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1006.903369] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d5abefd-375a-47a0-85a4-43f4e7373a06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.911379] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1006.911379] env[61947]: value = "task-1224840" [ 1006.911379] env[61947]: _type = "Task" [ 1006.911379] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.919382] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.997070] env[61947]: DEBUG nova.scheduler.client.report [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.053886] env[61947]: DEBUG nova.compute.manager [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Received event network-vif-plugged-a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.054217] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Acquiring lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.054343] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.054562] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.054678] env[61947]: DEBUG nova.compute.manager [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] No waiting events found dispatching network-vif-plugged-a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.054846] env[61947]: WARNING nova.compute.manager [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Received unexpected event network-vif-plugged-a5b4c8f3-144d-4a20-855b-41e891f9ccda for instance with vm_state building and task_state spawning. [ 1007.055018] env[61947]: DEBUG nova.compute.manager [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Received event network-changed-a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.055181] env[61947]: DEBUG nova.compute.manager [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Refreshing instance network info cache due to event network-changed-a5b4c8f3-144d-4a20-855b-41e891f9ccda. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1007.055366] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Acquiring lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.065669] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.080214] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.080688] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Instance network_info: |[{"id": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "address": "fa:16:3e:b1:6d:dd", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b4c8f3-14", "ovs_interfaceid": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.081383] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Acquired lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.081383] env[61947]: DEBUG nova.network.neutron [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Refreshing network info cache for port a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.083020] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:6d:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '438671d0-9468-4e44-84c1-4c0ebaa743e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5b4c8f3-144d-4a20-855b-41e891f9ccda', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.091724] env[61947]: DEBUG oslo.service.loopingcall [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.095087] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1007.095601] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c254293-0685-43c8-a5cd-fd311daeff9d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.119991] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.119991] env[61947]: value = "task-1224841" [ 1007.119991] env[61947]: _type = "Task" [ 1007.119991] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.130055] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224841, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.234324] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.265490] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.265817] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.266012] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.266232] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.266392] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.266582] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.266789] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.266962] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.267244] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.267444] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.267656] env[61947]: DEBUG nova.virt.hardware [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.268650] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c214ca1-066f-496e-b91b-2c268de678bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.277559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-528969b0-6172-490e-884b-6a50a9043820 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.283509] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba982e7-e453-4d01-b094-3efee5251954 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.361687] env[61947]: DEBUG nova.network.neutron [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Updated VIF entry in instance network info cache for port a5b4c8f3-144d-4a20-855b-41e891f9ccda. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.362168] env[61947]: DEBUG nova.network.neutron [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Updating instance_info_cache with network_info: [{"id": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "address": "fa:16:3e:b1:6d:dd", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b4c8f3-14", "ovs_interfaceid": "a5b4c8f3-144d-4a20-855b-41e891f9ccda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.425750] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224840, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.505095] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.505095] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.505693] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.356s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.505744] env[61947]: DEBUG nova.objects.instance [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'resources' on Instance uuid bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.527127] env[61947]: DEBUG nova.compute.manager [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Received event network-vif-plugged-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.527343] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] Acquiring lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.527629] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] Lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.527918] env[61947]: DEBUG oslo_concurrency.lockutils [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] Lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.527998] env[61947]: DEBUG nova.compute.manager [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] No waiting events found dispatching network-vif-plugged-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.528202] env[61947]: WARNING nova.compute.manager [req-f5b164a8-c740-4300-96a1-4dec55cf6a01 req-a0874e73-7593-4b80-b4eb-e187da539913 service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Received unexpected event network-vif-plugged-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a for instance with vm_state building and task_state spawning. [ 1007.567468] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.631135] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224841, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.688113] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Successfully updated port: 3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.865562] env[61947]: DEBUG oslo_concurrency.lockutils [req-edffa98d-67d9-4223-9ce1-88b657d4d23c req-50aca552-163c-4581-aeda-baad53499e13 service nova] Releasing lock "refresh_cache-8a4b81d3-d683-4517-806c-88b41ad73c48" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.922659] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625367} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.922935] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 1660df2a-b2c5-469f-b602-0014dd3656d5/1660df2a-b2c5-469f-b602-0014dd3656d5.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1007.923172] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.923433] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4198e635-cdfb-435e-80dc-30a7ff610649 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.931672] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1007.931672] env[61947]: value = "task-1224842" [ 1007.931672] env[61947]: _type = "Task" [ 1007.931672] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.940954] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.009755] env[61947]: DEBUG nova.compute.utils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.014913] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.014913] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1008.053677] env[61947]: DEBUG nova.policy [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6b5225bc6a74cae95357bc0ef03bbaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2367a725fa784cc9bbe918bcdf4b12e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1008.066711] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.132825] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224841, 'name': CreateVM_Task, 'duration_secs': 0.604878} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.132825] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1008.133338] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.133508] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.133821] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.136266] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa9d0134-b795-4857-bade-9727bac88fc0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.142214] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1008.142214] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a885b8-c2f9-d82b-8c0d-c6c609e1effc" [ 1008.142214] env[61947]: _type = "Task" [ 1008.142214] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.151486] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a885b8-c2f9-d82b-8c0d-c6c609e1effc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.192931] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.193097] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.193255] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.203060] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dcf746-ff86-4df5-b4e0-f941b397c93f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.213867] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f9a424-9bd7-4d10-bf3b-09b301d14add {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.247342] env[61947]: DEBUG nova.objects.instance [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.249515] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652b6662-5329-44fc-8d8e-5a51626150bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.258525] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56b444a-3064-4e19-afc3-dbf277c87501 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.275376] env[61947]: DEBUG nova.compute.provider_tree [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.353534] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Successfully created port: 8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.442474] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068558} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.443056] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.444017] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fec0be2-f8fd-4cd9-9e30-8b6ba16215bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.467389] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 1660df2a-b2c5-469f-b602-0014dd3656d5/1660df2a-b2c5-469f-b602-0014dd3656d5.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.467722] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c54bb37-20b8-4853-b29d-294fb16ee9a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.491531] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1008.491531] env[61947]: value = "task-1224843" [ 1008.491531] env[61947]: _type = "Task" [ 1008.491531] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.500431] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.515198] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.568995] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.653525] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a885b8-c2f9-d82b-8c0d-c6c609e1effc, 'name': SearchDatastore_Task, 'duration_secs': 0.019856} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.653900] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.654197] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.654513] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.654715] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.654932] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.655261] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-633e32af-9ad6-433d-9d01-999d50301307 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.665682] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.665904] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1008.666843] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d51bdcf2-bd8d-49ca-a31b-eef649c027ed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.672806] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1008.672806] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520ba63f-bf79-8a67-b0ae-204313a86b72" [ 1008.672806] env[61947]: _type = "Task" [ 1008.672806] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.681018] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520ba63f-bf79-8a67-b0ae-204313a86b72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.725765] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.756035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.756035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.756124] env[61947]: DEBUG nova.network.neutron [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.756279] env[61947]: DEBUG nova.objects.instance [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'info_cache' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.777977] env[61947]: DEBUG nova.scheduler.client.report [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.880681] env[61947]: DEBUG nova.network.neutron [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Updating instance_info_cache with network_info: [{"id": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "address": "fa:16:3e:db:90:c9", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e2fe97e-65", "ovs_interfaceid": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.001969] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224843, 'name': ReconfigVM_Task, 'duration_secs': 0.307477} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.002639] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 1660df2a-b2c5-469f-b602-0014dd3656d5/1660df2a-b2c5-469f-b602-0014dd3656d5.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.003359] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e5e2c44-4516-498c-8fba-4a47b88586b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.010762] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1009.010762] env[61947]: value = "task-1224844" [ 1009.010762] env[61947]: _type = "Task" [ 1009.010762] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.022427] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224844, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.068699] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.182908] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520ba63f-bf79-8a67-b0ae-204313a86b72, 'name': SearchDatastore_Task, 'duration_secs': 0.023403} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.183705] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-527b6b23-9aa2-4551-92df-69534da235a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.188785] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1009.188785] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cf73cb-1ef4-d522-7e02-14137c92e22c" [ 1009.188785] env[61947]: _type = "Task" [ 1009.188785] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.196819] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cf73cb-1ef4-d522-7e02-14137c92e22c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.262553] env[61947]: DEBUG nova.objects.base [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Object Instance<21bb0270-bc20-4ec1-9599-d676845b0dc7> lazy-loaded attributes: flavor,info_cache {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1009.282874] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.777s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.285164] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.137s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.286789] env[61947]: INFO nova.compute.claims [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.303821] env[61947]: INFO nova.scheduler.client.report [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted allocations for instance bd84b03e-ef6d-41e0-b48c-b8e8fc425a36 [ 1009.383488] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.383845] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Instance network_info: |[{"id": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "address": "fa:16:3e:db:90:c9", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e2fe97e-65", "ovs_interfaceid": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.384356] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:90:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '438671d0-9468-4e44-84c1-4c0ebaa743e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e2fe97e-65e9-40f3-a6f3-4008cd00d83a', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.392461] env[61947]: DEBUG oslo.service.loopingcall [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.392695] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1009.392930] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31eafca1-a53f-4697-96a3-e6d0c0bd59b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.413073] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.413073] env[61947]: value = "task-1224845" [ 1009.413073] env[61947]: _type = "Task" [ 1009.413073] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.421215] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224845, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.525963] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.528279] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224844, 'name': Rename_Task, 'duration_secs': 0.163454} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.532165] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1009.532165] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22059e6d-92aa-4398-9ac8-d4a618565a1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.537286] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1009.537286] env[61947]: value = "task-1224846" [ 1009.537286] env[61947]: _type = "Task" [ 1009.537286] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.548336] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.557015] env[61947]: DEBUG nova.compute.manager [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Received event network-changed-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.557015] env[61947]: DEBUG nova.compute.manager [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Refreshing instance network info cache due to event network-changed-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1009.557397] env[61947]: DEBUG oslo_concurrency.lockutils [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] Acquiring lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.557482] env[61947]: DEBUG oslo_concurrency.lockutils [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] Acquired lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.557701] env[61947]: DEBUG nova.network.neutron [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Refreshing network info cache for port 3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.561306] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.561886] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.561886] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.562198] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.562443] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.562688] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.563019] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.563281] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.563570] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.563847] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.564148] env[61947]: DEBUG nova.virt.hardware [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.565894] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226b0c3d-e2c0-4c1d-a5e5-93f11fdcf466 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.587804] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9c3586-a8fe-4381-ab44-a223b77623ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.592262] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.699343] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cf73cb-1ef4-d522-7e02-14137c92e22c, 'name': SearchDatastore_Task, 'duration_secs': 0.009687} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.699629] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.699935] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 8a4b81d3-d683-4517-806c-88b41ad73c48/8a4b81d3-d683-4517-806c-88b41ad73c48.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1009.700229] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3d6e94f-134a-4a84-ae47-9f1c9bdd6746 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.711221] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1009.711221] env[61947]: value = "task-1224847" [ 1009.711221] env[61947]: _type = "Task" [ 1009.711221] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.721544] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.811021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-dab36667-e15d-4651-a347-9ebe20b7cb7d tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "bd84b03e-ef6d-41e0-b48c-b8e8fc425a36" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.281s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.924400] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224845, 'name': CreateVM_Task, 'duration_secs': 0.361917} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.924661] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1009.925651] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.925809] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.926205] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.926524] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d58aa798-c037-4fc9-9a0e-69102ed0cbf2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.938181] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1009.938181] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f490a4-3a58-a558-0fbb-96fbe94a3519" [ 1009.938181] env[61947]: _type = "Task" [ 1009.938181] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.955088] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f490a4-3a58-a558-0fbb-96fbe94a3519, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.049560] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224846, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.076453] env[61947]: DEBUG oslo_vmware.api [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224835, 'name': ReconfigVM_Task, 'duration_secs': 5.79509} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.076745] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.076967] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Reconfigured VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1010.134638] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Successfully updated port: 8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.226117] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473435} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.226461] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 8a4b81d3-d683-4517-806c-88b41ad73c48/8a4b81d3-d683-4517-806c-88b41ad73c48.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1010.226701] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.227009] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7b3e17e-c6e6-480e-aea8-e4c76d64fb99 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.235547] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1010.235547] env[61947]: value = "task-1224848" [ 1010.235547] env[61947]: _type = "Task" [ 1010.235547] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.244301] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224848, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.414354] env[61947]: DEBUG nova.network.neutron [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Updated VIF entry in instance network info cache for port 3e2fe97e-65e9-40f3-a6f3-4008cd00d83a. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.415163] env[61947]: DEBUG nova.network.neutron [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Updating instance_info_cache with network_info: [{"id": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "address": "fa:16:3e:db:90:c9", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e2fe97e-65", "ovs_interfaceid": "3e2fe97e-65e9-40f3-a6f3-4008cd00d83a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.461187] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52f490a4-3a58-a558-0fbb-96fbe94a3519, 'name': SearchDatastore_Task, 'duration_secs': 0.055625} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.461187] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.461187] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.461187] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.461187] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.461187] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.462025] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55cb137e-9f27-4a99-900d-8ddf3cad7a72 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.479180] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.479457] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.480672] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2f73cd-ada8-4c26-8af1-50a2c6ac40cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.489958] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1010.489958] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b2ffe8-7340-72b9-7f3d-7fc43a253d44" [ 1010.489958] env[61947]: _type = "Task" [ 1010.489958] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.499264] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b2ffe8-7340-72b9-7f3d-7fc43a253d44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.513301] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecc0fbb-5a4f-4537-88db-82300e080e91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.523521] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612435df-5a89-46ac-8d0d-c1983ca0b387 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.560367] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233ffe51-077e-4286-b6d7-7094875b6972 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.571816] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3521c0c6-56ba-460d-906a-41e10f2782e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.575427] env[61947]: DEBUG oslo_vmware.api [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224846, 'name': PowerOnVM_Task, 'duration_secs': 0.811221} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.575749] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1010.575992] env[61947]: INFO nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1010.576230] env[61947]: DEBUG nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.577325] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ede927-2384-4dea-bdf5-01650e2b8155 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.589681] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.638049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.638049] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.638049] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.663404] env[61947]: DEBUG nova.network.neutron [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.747318] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224848, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096636} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.747318] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.748159] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b33165-c3ec-4803-9d23-502931167073 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.775847] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 8a4b81d3-d683-4517-806c-88b41ad73c48/8a4b81d3-d683-4517-806c-88b41ad73c48.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.776183] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96dc6e22-018f-4b55-861c-e7125622beb9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.799499] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1010.799499] env[61947]: value = "task-1224849" [ 1010.799499] env[61947]: _type = "Task" [ 1010.799499] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.809198] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.919946] env[61947]: DEBUG oslo_concurrency.lockutils [req-e27ad9dd-9e54-44e8-8f7c-3405a32feaee req-d30a5d45-33ca-4ca1-9ad6-c78f2f29c7ee service nova] Releasing lock "refresh_cache-4e0a8c79-9445-4474-b516-e697e1854dea" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.002158] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b2ffe8-7340-72b9-7f3d-7fc43a253d44, 'name': SearchDatastore_Task, 'duration_secs': 0.079166} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.002920] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0dad07d-892c-41f9-86c6-68d2188cf712 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.009060] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1011.009060] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ad5cc4-7ccb-bea8-127b-a276cc0c1f4c" [ 1011.009060] env[61947]: _type = "Task" [ 1011.009060] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.017490] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ad5cc4-7ccb-bea8-127b-a276cc0c1f4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.106682] env[61947]: INFO nova.compute.manager [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Took 16.64 seconds to build instance. [ 1011.114590] env[61947]: ERROR nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [req-0e3a330a-46a7-4fb7-8595-46b18f4f444a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0e3a330a-46a7-4fb7-8595-46b18f4f444a"}]} [ 1011.130535] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1011.146992] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1011.147354] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.163194] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1011.166136] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.178452] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1011.192307] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1011.315239] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224849, 'name': ReconfigVM_Task, 'duration_secs': 0.310606} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.315536] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 8a4b81d3-d683-4517-806c-88b41ad73c48/8a4b81d3-d683-4517-806c-88b41ad73c48.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.316171] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e0da02c-3332-4cfe-80c9-e33aa84b791b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.323781] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1011.323781] env[61947]: value = "task-1224850" [ 1011.323781] env[61947]: _type = "Task" [ 1011.323781] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.335406] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224850, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.371314] env[61947]: DEBUG nova.network.neutron [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.395069] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4998426-9ead-4e27-a874-301494a24167 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.404488] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a60b344-488c-4dd4-a2d7-2c72dcec1e36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.438420] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2ea86e-067f-4b2a-b4d0-3f2e8a9b5ba1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.537240] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59fb1a5-1aca-46a5-b4c6-3d75d3b137cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.537240] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.537240] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ad5cc4-7ccb-bea8-127b-a276cc0c1f4c, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.537240] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.537240] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4e0a8c79-9445-4474-b516-e697e1854dea/4e0a8c79-9445-4474-b516-e697e1854dea.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.537240] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24d6508b-7811-46bc-8302-e1f018475737 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.537240] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1011.537240] env[61947]: value = "task-1224851" [ 1011.537240] env[61947]: _type = "Task" [ 1011.537240] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.539167] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.609534] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c3b8e3aa-e0a4-438f-9ba8-b5c17dda3183 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.155s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.669579] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1011.670155] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3eeab8c7-9e08-4703-affa-4268a5e52e81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.678536] env[61947]: DEBUG oslo_vmware.api [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1011.678536] env[61947]: value = "task-1224852" [ 1011.678536] env[61947]: _type = "Task" [ 1011.678536] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.688189] env[61947]: DEBUG oslo_vmware.api [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.760094] env[61947]: DEBUG nova.compute.manager [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Received event network-vif-plugged-8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.760579] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.760579] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.760877] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.760877] env[61947]: DEBUG nova.compute.manager [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] No waiting events found dispatching network-vif-plugged-8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.761046] env[61947]: WARNING nova.compute.manager [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Received unexpected event network-vif-plugged-8e500219-2108-4201-adbc-0378294780ef for instance with vm_state building and task_state spawning. [ 1011.761219] env[61947]: DEBUG nova.compute.manager [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Received event network-changed-8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.761382] env[61947]: DEBUG nova.compute.manager [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Refreshing instance network info cache due to event network-changed-8e500219-2108-4201-adbc-0378294780ef. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1011.761553] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Acquiring lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.797404] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.797646] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.797868] env[61947]: DEBUG nova.network.neutron [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.837927] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224850, 'name': Rename_Task, 'duration_secs': 0.170225} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.837927] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1011.837927] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-269b391f-27e0-401c-924d-4a635a212699 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.844042] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1011.844042] env[61947]: value = "task-1224853" [ 1011.844042] env[61947]: _type = "Task" [ 1011.844042] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.854129] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.873966] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.874445] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Instance network_info: |[{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1011.874777] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Acquired lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.875040] env[61947]: DEBUG nova.network.neutron [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Refreshing network info cache for port 8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.877222] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:2d:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e500219-2108-4201-adbc-0378294780ef', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.884714] env[61947]: DEBUG oslo.service.loopingcall [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.886301] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1011.886563] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8374daf0-9ffb-4aab-99cc-1f9a1ab345f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.909675] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.909675] env[61947]: value = "task-1224854" [ 1011.909675] env[61947]: _type = "Task" [ 1011.909675] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.919121] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224854, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.983876] env[61947]: ERROR nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [req-2ba0feb7-9e3d-4e24-9285-54779fc058c5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ba0feb7-9e3d-4e24-9285-54779fc058c5"}]} [ 1012.003784] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1012.016197] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1012.016686] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.028933] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1012.042110] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.049981] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1012.188899] env[61947]: DEBUG oslo_vmware.api [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224852, 'name': PowerOnVM_Task, 'duration_secs': 0.461493} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.189447] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1012.189881] env[61947]: DEBUG nova.compute.manager [None req-0bcfedb9-f03b-4a42-bfcf-629629caec4e tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.190964] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2de243-f20f-4c2f-9f32-4ad1e06f7148 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.241706] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8594e6-aec2-4cd3-bf12-8f81db258784 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.252686] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141af51c-acf5-4a70-817d-4425d095d04f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.288953] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1ded06-72c5-415b-a66e-7e07c3956935 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.296280] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.296518] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.304620] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0013c6-b148-431c-a64a-8467e2b5693a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.320434] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1012.320886] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.322173] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1012.355770] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224853, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.420253] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224854, 'name': CreateVM_Task, 'duration_secs': 0.485462} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.420424] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1012.421117] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.421279] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.421694] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.422295] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d06bb0a2-df47-4376-8d19-067220af5c8a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.427730] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1012.427730] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52baea36-092e-ca7a-9983-fc73ce582348" [ 1012.427730] env[61947]: _type = "Task" [ 1012.427730] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.438763] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52baea36-092e-ca7a-9983-fc73ce582348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.542181] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.759513] env[61947]: INFO nova.network.neutron [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Port 76c0251c-9621-4df0-83f5-b4a19d1f5da9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1012.759926] env[61947]: DEBUG nova.network.neutron [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.804782] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1012.819198] env[61947]: DEBUG nova.network.neutron [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updated VIF entry in instance network info cache for port 8e500219-2108-4201-adbc-0378294780ef. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.819762] env[61947]: DEBUG nova.network.neutron [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.834908] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1012.835213] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1012.835411] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1012.864101] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224853, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.875720] env[61947]: DEBUG nova.scheduler.client.report [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 127 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1012.876152] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 127 to 128 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1012.876433] env[61947]: DEBUG nova.compute.provider_tree [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.940939] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52baea36-092e-ca7a-9983-fc73ce582348, 'name': SearchDatastore_Task, 'duration_secs': 0.01086} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.941330] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.941718] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.942078] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.942569] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.942777] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.943038] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d6127c4-eaeb-41d5-8ebe-b5fa091e13e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.953264] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.953899] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1012.954326] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1d2f4b-9d94-449b-bf07-1aa8ab02529e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.960328] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1012.960328] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52daa433-b6ec-fa6e-868d-0abfd590edb1" [ 1012.960328] env[61947]: _type = "Task" [ 1012.960328] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.969734] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52daa433-b6ec-fa6e-868d-0abfd590edb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.043782] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224851, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.410876} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.044018] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 4e0a8c79-9445-4474-b516-e697e1854dea/4e0a8c79-9445-4474-b516-e697e1854dea.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1013.044142] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.044358] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ac55367-59b6-4c00-b1af-fd0cdd39246f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.051073] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1013.051073] env[61947]: value = "task-1224855" [ 1013.051073] env[61947]: _type = "Task" [ 1013.051073] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.060234] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.263490] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.279653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.279916] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.280286] env[61947]: DEBUG nova.objects.instance [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'flavor' on Instance uuid e137f21e-766d-4b20-9d92-5d1907e2baa3 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.323086] env[61947]: DEBUG oslo_concurrency.lockutils [req-4aba1d0f-ad4e-4124-8955-22cdb39d8868 req-9616382a-4ada-43c0-9f5e-b5ce6e7b08a5 service nova] Releasing lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.327132] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.339838] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1013.339982] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1013.340124] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1013.340250] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1013.356129] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224853, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.381634] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.381875] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.382040] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1013.382206] env[61947]: DEBUG nova.objects.instance [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lazy-loading 'info_cache' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.383964] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.099s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.384463] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.387153] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.052s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.389253] env[61947]: INFO nova.compute.claims [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.471852] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52daa433-b6ec-fa6e-868d-0abfd590edb1, 'name': SearchDatastore_Task, 'duration_secs': 0.010299} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.472715] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bc77dc7-044e-4647-a2ea-5068d4d9dde1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.478193] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1013.478193] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528721aa-775b-2458-7050-38d0c9016978" [ 1013.478193] env[61947]: _type = "Task" [ 1013.478193] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.485840] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528721aa-775b-2458-7050-38d0c9016978, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.561156] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102344} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.561467] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.562253] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da287b4-9fb5-4149-9d6e-4002d7ed221c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.585148] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 4e0a8c79-9445-4474-b516-e697e1854dea/4e0a8c79-9445-4474-b516-e697e1854dea.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.585581] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac66ee9c-e81e-4043-a3b3-cb38479ddf80 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.605072] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1013.605072] env[61947]: value = "task-1224856" [ 1013.605072] env[61947]: _type = "Task" [ 1013.605072] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.612952] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.767531] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2ec7f21e-8656-4b38-9b8f-b35311927919 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-76c0251c-9621-4df0-83f5-b4a19d1f5da9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.310s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.807888] env[61947]: DEBUG nova.compute.manager [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.808108] env[61947]: DEBUG nova.compute.manager [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing instance network info cache due to event network-changed-c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1013.808358] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Acquiring lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.808480] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Acquired lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.808722] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Refreshing network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.856107] env[61947]: DEBUG oslo_vmware.api [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224853, 'name': PowerOnVM_Task, 'duration_secs': 1.556858} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.856393] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1013.856600] env[61947]: INFO nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1013.856782] env[61947]: DEBUG nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.857587] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4bf4dc-409e-494f-8732-b70a93247891 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.878706] env[61947]: DEBUG nova.objects.instance [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'pci_requests' on Instance uuid e137f21e-766d-4b20-9d92-5d1907e2baa3 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.894084] env[61947]: DEBUG nova.compute.utils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.897134] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.897302] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.935458] env[61947]: DEBUG nova.policy [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '720d781b9db640368530ccaeaeebf993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c530a8353f724243982b78eaa601b131', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1013.990073] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528721aa-775b-2458-7050-38d0c9016978, 'name': SearchDatastore_Task, 'duration_secs': 0.04533} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.990073] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.990073] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1013.990291] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2c430f7-aebe-4c1c-83e2-03a4ba485594 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.998334] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1013.998334] env[61947]: value = "task-1224857" [ 1013.998334] env[61947]: _type = "Task" [ 1013.998334] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.006567] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224857, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.116779] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.235776] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Successfully created port: aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.374852] env[61947]: INFO nova.compute.manager [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Took 19.66 seconds to build instance. [ 1014.382234] env[61947]: DEBUG nova.objects.base [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.382548] env[61947]: DEBUG nova.network.neutron [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.403613] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.517666] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224857, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.554047] env[61947]: DEBUG nova.policy [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ed05c3b298c417c88af3c6e8786fb19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb7a249773754feaa2ff8a3b9d5bea54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1014.620033] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224856, 'name': ReconfigVM_Task, 'duration_secs': 0.771286} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.620527] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 4e0a8c79-9445-4474-b516-e697e1854dea/4e0a8c79-9445-4474-b516-e697e1854dea.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.621253] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f24dfcb-ff7c-4b10-ba00-a3fe95694f68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.628546] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1014.628546] env[61947]: value = "task-1224858" [ 1014.628546] env[61947]: _type = "Task" [ 1014.628546] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.639599] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224858, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.707377] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f75fa4-e6dd-4e60-b601-abaf45a55f3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.717464] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1127b6c1-1a0f-445c-a837-6e2e0ad1c02b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.751291] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a0525f-02d4-43f9-978f-4bfc012a4d87 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.766865] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef28f1a-579b-4f36-8d72-01e66ca4f604 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.781306] env[61947]: DEBUG nova.compute.provider_tree [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.785840] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updated VIF entry in instance network info cache for port c32a51a8-436f-4d1d-abb7-452a152efd91. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.785840] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [{"id": "c32a51a8-436f-4d1d-abb7-452a152efd91", "address": "fa:16:3e:1a:87:3e", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32a51a8-43", "ovs_interfaceid": "c32a51a8-436f-4d1d-abb7-452a152efd91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.877257] env[61947]: DEBUG oslo_concurrency.lockutils [None req-65a042cd-bed8-47be-be65-77c7cce9e14e tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.172s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.008848] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224857, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.126027] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [{"id": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "address": "fa:16:3e:c6:b6:3f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b7626fb-f4", "ovs_interfaceid": "0b7626fb-f420-4ad3-bea5-09a0daafed10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.140935] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224858, 'name': Rename_Task, 'duration_secs': 0.155978} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.141477] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1015.141477] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f96a3d49-96af-426f-b73d-5fde5de59a72 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.149017] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1015.149017] env[61947]: value = "task-1224859" [ 1015.149017] env[61947]: _type = "Task" [ 1015.149017] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.157838] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.285114] env[61947]: DEBUG nova.scheduler.client.report [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.288215] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Releasing lock "refresh_cache-fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.288453] env[61947]: DEBUG nova.compute.manager [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.288626] env[61947]: DEBUG nova.compute.manager [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing instance network info cache due to event network-changed-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1015.288831] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.288976] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.289153] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.415132] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.440505] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.440795] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.440959] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.441157] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.441308] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.441463] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.441676] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.441839] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.442022] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.442193] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.442369] env[61947]: DEBUG nova.virt.hardware [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.443252] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aab9b9b-d227-44cb-aac8-cb4ca5058423 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.451820] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3196ac4b-4427-4354-869f-52a41a532e79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.507970] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224857, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.629630] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-21bb0270-bc20-4ec1-9599-d676845b0dc7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.629867] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1015.630098] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630261] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630442] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630558] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630703] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630847] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.630974] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1015.631130] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1015.659947] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.791892] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.792702] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1015.799665] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.367s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.801707] env[61947]: INFO nova.compute.claims [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.010339] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224857, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.54464} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.010536] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1016.014019] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.014019] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dac81ffd-1fc2-4558-a6b8-aac8cc354dc7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.018099] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1016.018099] env[61947]: value = "task-1224860" [ 1016.018099] env[61947]: _type = "Task" [ 1016.018099] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.026461] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.134722] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.160787] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.221112] env[61947]: DEBUG nova.compute.manager [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Received event network-vif-plugged-aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.221359] env[61947]: DEBUG oslo_concurrency.lockutils [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] Acquiring lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.221581] env[61947]: DEBUG oslo_concurrency.lockutils [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.221757] env[61947]: DEBUG oslo_concurrency.lockutils [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.221930] env[61947]: DEBUG nova.compute.manager [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] No waiting events found dispatching network-vif-plugged-aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.222118] env[61947]: WARNING nova.compute.manager [req-602f3de0-d3ac-4015-a031-9cf7394e4240 req-c4c2798f-4d27-4e3c-8e3a-ae8e3b83e53b service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Received unexpected event network-vif-plugged-aea546b4-1737-4566-8328-b0dcaca4e4e9 for instance with vm_state building and task_state spawning. [ 1016.257931] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updated VIF entry in instance network info cache for port 1e4942ef-a8cf-44ec-bc80-58f9ae5894a5. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.258337] env[61947]: DEBUG nova.network.neutron [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.301060] env[61947]: DEBUG nova.compute.utils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.302489] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.302664] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1016.395504] env[61947]: DEBUG nova.policy [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a0420988621c4dd39cfbc99db4b9719e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d76ebc8875f4b03bc768320b44b11d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1016.404779] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Successfully updated port: aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.488366] env[61947]: DEBUG nova.network.neutron [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Successfully updated port: 76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.528270] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.199893} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.529053] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.530290] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b3d652-a76e-414c-9fdd-86835d5fe4f1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.553667] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.553983] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e654442e-3421-48b3-bfd6-2f63deab9d91 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.575046] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1016.575046] env[61947]: value = "task-1224861" [ 1016.575046] env[61947]: _type = "Task" [ 1016.575046] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.583818] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224861, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.661182] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224859, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.676628] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Successfully created port: d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1016.761544] env[61947]: DEBUG oslo_concurrency.lockutils [req-89b2150b-6788-411f-9937-50dde2ce0f77 req-2e3e3e11-900f-45b4-816c-f04c327a35b1 service nova] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.805944] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1016.908038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.908198] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquired lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.908353] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.980275] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de8386d-2a0a-4e1b-a394-a95b48f468e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.988486] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4847af9-d170-4855-9c49-db1eb31b3da4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.992152] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.992375] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.992571] env[61947]: DEBUG nova.network.neutron [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.022841] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb9b6c7-f07f-4ff9-83c3-8be15bc40d7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.031194] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f09579-ab58-4a28-9b43-9331a61ce5b9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.045096] env[61947]: DEBUG nova.compute.provider_tree [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.084997] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224861, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.161479] env[61947]: DEBUG oslo_vmware.api [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224859, 'name': PowerOnVM_Task, 'duration_secs': 1.592836} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.161777] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.161985] env[61947]: INFO nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Took 9.93 seconds to spawn the instance on the hypervisor. [ 1017.162189] env[61947]: DEBUG nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.163051] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295782d7-607b-40fc-8bb6-5965b6174ef7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.441543] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1017.538200] env[61947]: WARNING nova.network.neutron [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] 720a7469-85cf-4e25-82f3-b574730aae1f already exists in list: networks containing: ['720a7469-85cf-4e25-82f3-b574730aae1f']. ignoring it [ 1017.547971] env[61947]: DEBUG nova.scheduler.client.report [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.585801] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224861, 'name': ReconfigVM_Task, 'duration_secs': 0.855536} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.586098] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.586704] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25b11417-b869-47f9-944e-d07f2530e2ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.595556] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1017.595556] env[61947]: value = "task-1224862" [ 1017.595556] env[61947]: _type = "Task" [ 1017.595556] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.605710] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224862, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.662844] env[61947]: DEBUG nova.network.neutron [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updating instance_info_cache with network_info: [{"id": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "address": "fa:16:3e:ba:8e:f0", "network": {"id": "247751fc-21e1-4170-b756-d54244d27b23", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1763905409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c530a8353f724243982b78eaa601b131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaea546b4-17", "ovs_interfaceid": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.680827] env[61947]: INFO nova.compute.manager [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Took 22.56 seconds to build instance. [ 1017.818042] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1017.843968] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.843968] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.843968] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.843968] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.843968] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.844314] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.844355] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.845027] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.845027] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.845027] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.845230] env[61947]: DEBUG nova.virt.hardware [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.846086] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4279b43-7c92-4a47-92dc-b2ea2bcc0117 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.854825] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86928052-4784-42e6-9862-85fa11b36df1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.873744] env[61947]: DEBUG nova.network.neutron [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "address": "fa:16:3e:85:91:f7", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76c0251c-96", "ovs_interfaceid": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.053508] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.054082] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.056735] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.730s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.058650] env[61947]: INFO nova.compute.claims [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1018.107426] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224862, 'name': Rename_Task, 'duration_secs': 0.161472} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.107426] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1018.107664] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9123313b-8ba3-491a-a95b-cb5d7449bf6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.116221] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1018.116221] env[61947]: value = "task-1224863" [ 1018.116221] env[61947]: _type = "Task" [ 1018.116221] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.126290] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.165952] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Releasing lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.165952] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Instance network_info: |[{"id": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "address": "fa:16:3e:ba:8e:f0", "network": {"id": "247751fc-21e1-4170-b756-d54244d27b23", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1763905409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c530a8353f724243982b78eaa601b131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaea546b4-17", "ovs_interfaceid": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.171105] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:8e:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f0ef5aba-bd9a-42ff-a1a0-5e763986d70a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aea546b4-1737-4566-8328-b0dcaca4e4e9', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.177048] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Creating folder: Project (c530a8353f724243982b78eaa601b131). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.178010] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Successfully updated port: d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.181050] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e41303eb-3235-443e-a047-ebf6981fe15d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.185611] env[61947]: DEBUG oslo_concurrency.lockutils [None req-474b02f7-09d8-44f8-a4a7-0b7e49e0a869 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.071s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.197173] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Created folder: Project (c530a8353f724243982b78eaa601b131) in parent group-v264556. [ 1018.197457] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Creating folder: Instances. Parent ref: group-v264723. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.197751] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75aa8cd8-7f44-4c22-86b1-68d04dd04542 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.211079] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Created folder: Instances in parent group-v264723. [ 1018.211428] env[61947]: DEBUG oslo.service.loopingcall [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.211686] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.212035] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0d6cc59-6087-457e-8f43-56c3d5d6f30f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.235155] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.235155] env[61947]: value = "task-1224866" [ 1018.235155] env[61947]: _type = "Task" [ 1018.235155] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.243964] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224866, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.350123] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.350123] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.350824] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.351090] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.351611] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] No waiting events found dispatching network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.351903] env[61947]: WARNING nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received unexpected event network-vif-plugged-76c0251c-9621-4df0-83f5-b4a19d1f5da9 for instance with vm_state active and task_state None. [ 1018.352146] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Received event network-changed-aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.352356] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Refreshing instance network info cache due to event network-changed-aea546b4-1737-4566-8328-b0dcaca4e4e9. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.352588] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquiring lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.352762] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquired lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.352958] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Refreshing network info cache for port aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.376465] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.377172] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.377353] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.378479] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30a7926-1aa7-41d6-bedd-897ae2fd88b2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.397239] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.397492] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.397706] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.397964] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.398162] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.398347] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.398566] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.398755] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.398945] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.399155] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.399369] env[61947]: DEBUG nova.virt.hardware [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.405711] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfiguring VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1018.406097] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34983670-6d5b-4627-bfe7-b3fd8052f5a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.425013] env[61947]: DEBUG oslo_vmware.api [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1018.425013] env[61947]: value = "task-1224867" [ 1018.425013] env[61947]: _type = "Task" [ 1018.425013] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.432572] env[61947]: DEBUG oslo_vmware.api [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224867, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.563356] env[61947]: DEBUG nova.compute.utils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.567356] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1018.567490] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1018.614727] env[61947]: DEBUG nova.policy [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7674313a124488d92ced546d42d583e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c07ddc4be8d4411a1762897ecfd89d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1018.628385] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224863, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.680880] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.680997] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.681173] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.745986] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224866, 'name': CreateVM_Task, 'duration_secs': 0.419283} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.746298] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.747036] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.747281] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.747670] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.747966] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a73add2-bef4-42ca-89dc-ccbe7b62576b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.753519] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1018.753519] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524b738e-df32-6228-ffbe-9983e8b05145" [ 1018.753519] env[61947]: _type = "Task" [ 1018.753519] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.763371] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524b738e-df32-6228-ffbe-9983e8b05145, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.898196] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Successfully created port: 6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.935635] env[61947]: DEBUG oslo_vmware.api [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224867, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.073472] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.130028] env[61947]: DEBUG oslo_vmware.api [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224863, 'name': PowerOnVM_Task, 'duration_secs': 0.765025} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.132729] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.132923] env[61947]: INFO nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Took 9.61 seconds to spawn the instance on the hypervisor. [ 1019.133128] env[61947]: DEBUG nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.134184] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca7e0a5-447e-4307-b0cd-d6d9393ee99e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.252336] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.273410] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524b738e-df32-6228-ffbe-9983e8b05145, 'name': SearchDatastore_Task, 'duration_secs': 0.043728} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.273742] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.274014] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.274290] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.274787] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.274985] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.275662] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abede278-d7c2-46b1-ab84-f794017be3d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.289596] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.289867] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.293621] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2d94940-4e3b-44f9-8ee8-d46a54dec2e2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.302359] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1019.302359] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cc70f7-5c05-07d2-cdc3-2e5ccaa33043" [ 1019.302359] env[61947]: _type = "Task" [ 1019.302359] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.311218] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cc70f7-5c05-07d2-cdc3-2e5ccaa33043, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.327876] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd6734f-d600-4794-93f0-0bb6feadbe11 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.338466] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1e6bff-5529-4822-bfa6-6cec8e834c2d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.383985] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3127fa08-5529-49e5-b414-c80eb9fd5e2a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.394117] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d736d21f-0d09-4105-9b8d-78d7e5e2216a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410924] env[61947]: DEBUG nova.compute.provider_tree [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.437379] env[61947]: DEBUG oslo_vmware.api [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224867, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.543039] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updated VIF entry in instance network info cache for port aea546b4-1737-4566-8328-b0dcaca4e4e9. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.543574] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updating instance_info_cache with network_info: [{"id": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "address": "fa:16:3e:ba:8e:f0", "network": {"id": "247751fc-21e1-4170-b756-d54244d27b23", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1763905409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c530a8353f724243982b78eaa601b131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaea546b4-17", "ovs_interfaceid": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.613789] env[61947]: DEBUG nova.network.neutron [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Updating instance_info_cache with network_info: [{"id": "d4f4f476-5bcf-482f-9c37-5436907cd763", "address": "fa:16:3e:96:4c:2c", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4f4f476-5b", "ovs_interfaceid": "d4f4f476-5bcf-482f-9c37-5436907cd763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.653216] env[61947]: INFO nova.compute.manager [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Took 22.22 seconds to build instance. [ 1019.813709] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52cc70f7-5c05-07d2-cdc3-2e5ccaa33043, 'name': SearchDatastore_Task, 'duration_secs': 0.01942} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.815052] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0eafb46-53f7-4be9-9d7f-b8a0db130217 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.820267] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1019.820267] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52038a0e-cbf6-da48-9ee6-f115040e0efa" [ 1019.820267] env[61947]: _type = "Task" [ 1019.820267] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.829511] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52038a0e-cbf6-da48-9ee6-f115040e0efa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.914702] env[61947]: DEBUG nova.scheduler.client.report [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.938227] env[61947]: DEBUG oslo_vmware.api [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224867, 'name': ReconfigVM_Task, 'duration_secs': 1.114817} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.938863] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.939146] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfigured VM to attach interface {{(pid=61947) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1020.047087] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Releasing lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.047337] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-changed-76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.047659] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing instance network info cache due to event network-changed-76c0251c-9621-4df0-83f5-b4a19d1f5da9. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.047902] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.048067] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.048243] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Refreshing network info cache for port 76c0251c-9621-4df0-83f5-b4a19d1f5da9 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.087912] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.112500] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.112784] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.112973] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.113180] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.113334] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.113486] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.113697] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.113866] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.114048] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.114233] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.114417] env[61947]: DEBUG nova.virt.hardware [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.115311] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc75f5f-8e1e-4b41-b09f-570381f1882e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.118064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.118356] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Instance network_info: |[{"id": "d4f4f476-5bcf-482f-9c37-5436907cd763", "address": "fa:16:3e:96:4c:2c", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4f4f476-5b", "ovs_interfaceid": "d4f4f476-5bcf-482f-9c37-5436907cd763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.118730] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:4c:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4f4f476-5bcf-482f-9c37-5436907cd763', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.126296] env[61947]: DEBUG oslo.service.loopingcall [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.127653] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1020.127801] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a021c88-d0e2-476d-b425-222f337dd26a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.144987] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91736d0-5feb-4003-8b41-f415f3f3cbb4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.149546] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.149546] env[61947]: value = "task-1224868" [ 1020.149546] env[61947]: _type = "Task" [ 1020.149546] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.161467] env[61947]: DEBUG oslo_concurrency.lockutils [None req-b7384ae0-93aa-407a-be11-3baeb8468ad7 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.740s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.166286] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224868, 'name': CreateVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.330735] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52038a0e-cbf6-da48-9ee6-f115040e0efa, 'name': SearchDatastore_Task, 'duration_secs': 0.012695} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.331089] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.331366] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 068c9c5e-f7e4-4a5e-896d-a5c348949e07/068c9c5e-f7e4-4a5e-896d-a5c348949e07.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1020.331628] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32ec830c-21e4-4081-b912-20e59c628226 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.337683] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1020.337683] env[61947]: value = "task-1224869" [ 1020.337683] env[61947]: _type = "Task" [ 1020.337683] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.345755] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.420386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.421070] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1020.424061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.289s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.424263] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.425257] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1020.425841] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f7cd26-2664-4ebb-9411-63288fb378b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.434033] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a42418-5490-4b3f-bf99-99763d123cb1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.449206] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2aa71dad-e251-47dd-908b-4fa4267a2c8b tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.168s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.451180] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcc094e-b3ff-4339-81a6-5d661293b44c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.457179] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a991d30-01c5-4f41-bb0b-0b525936d1b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.486804] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180611MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1020.486996] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.487188] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.661520] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224868, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.848585] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224869, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.854470] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updated VIF entry in instance network info cache for port 76c0251c-9621-4df0-83f5-b4a19d1f5da9. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.854694] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "address": "fa:16:3e:85:91:f7", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76c0251c-96", "ovs_interfaceid": "76c0251c-9621-4df0-83f5-b4a19d1f5da9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.927565] env[61947]: DEBUG nova.compute.utils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.929276] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1020.929461] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.968610] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.968867] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.003766] env[61947]: DEBUG nova.policy [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '897eb27f633a4a65bd500448ede60c1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7901e6bbfd1e4832a556d1a2b58d9850', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1021.160059] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224868, 'name': CreateVM_Task, 'duration_secs': 0.55153} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.160297] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1021.160976] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.161165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.161506] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.161760] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfb19c0e-3b8e-4a09-8a1b-561370ec8f6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.166686] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1021.166686] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52553e52-0376-47ad-bf07-7958ff5bd169" [ 1021.166686] env[61947]: _type = "Task" [ 1021.166686] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.174902] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52553e52-0376-47ad-bf07-7958ff5bd169, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.185367] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Successfully updated port: 6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.348172] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706096} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.348454] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 068c9c5e-f7e4-4a5e-896d-a5c348949e07/068c9c5e-f7e4-4a5e-896d-a5c348949e07.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.348684] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.348953] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20b4f279-8318-44a6-8446-61d06d58b3a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.355306] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1021.355306] env[61947]: value = "task-1224870" [ 1021.355306] env[61947]: _type = "Task" [ 1021.355306] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.358694] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.358948] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Received event network-vif-plugged-d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.359158] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquiring lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.359388] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.359559] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.359765] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] No waiting events found dispatching network-vif-plugged-d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.359977] env[61947]: WARNING nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Received unexpected event network-vif-plugged-d4f4f476-5bcf-482f-9c37-5436907cd763 for instance with vm_state building and task_state spawning. [ 1021.360170] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Received event network-changed-d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.360333] env[61947]: DEBUG nova.compute.manager [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Refreshing instance network info cache due to event network-changed-d4f4f476-5bcf-482f-9c37-5436907cd763. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1021.360523] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquiring lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.360722] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Acquired lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.360914] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Refreshing network info cache for port d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.366666] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.433153] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1021.473282] env[61947]: DEBUG nova.compute.utils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.477239] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Successfully created port: 7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.509215] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Skipping migration as instance is neither resizing nor live-migrating. {{(pid=61947) _update_usage_from_migrations /opt/stack/nova/nova/compute/resource_tracker.py:1563}} [ 1021.526664] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 21bb0270-bc20-4ec1-9599-d676845b0dc7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.526822] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance fcd2e134-4d5f-43d5-a09c-55a49f8ce48b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.526950] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance e137f21e-766d-4b20-9d92-5d1907e2baa3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527086] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 9378657d-dc54-47e2-b178-1ed3f9c952f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527209] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8351ae98-2d33-4f76-a3c1-937738905761 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527323] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 1660df2a-b2c5-469f-b602-0014dd3656d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527435] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8a4b81d3-d683-4517-806c-88b41ad73c48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527553] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 4e0a8c79-9445-4474-b516-e697e1854dea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527679] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 068c9c5e-f7e4-4a5e-896d-a5c348949e07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527789] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance a56bb67b-93e4-4ec1-a3c6-b9172a4ed617 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.527896] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance d68e1fc3-0129-432b-adce-30773f5bd6ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.528008] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 1faa141b-ddf6-4fc3-bb83-e77efa022f81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.528124] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Migration bdf8ba91-ccf6-4bc1-ab8c-7bc388fd7648 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1021.528231] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 0da87f27-f431-4ebf-ac5e-571605481627 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1021.528443] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1021.528581] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1021.649656] env[61947]: DEBUG nova.compute.manager [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Received event network-vif-plugged-6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.649929] env[61947]: DEBUG oslo_concurrency.lockutils [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] Acquiring lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.650206] env[61947]: DEBUG oslo_concurrency.lockutils [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.650381] env[61947]: DEBUG oslo_concurrency.lockutils [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.650555] env[61947]: DEBUG nova.compute.manager [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] No waiting events found dispatching network-vif-plugged-6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.650729] env[61947]: WARNING nova.compute.manager [req-89d634d9-c120-4327-b0ab-b1d885df05e7 req-0cccebdd-cb3c-4104-b830-31366e9f784b service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Received unexpected event network-vif-plugged-6c189068-6cb8-4da8-8db6-a717706e10fa for instance with vm_state building and task_state spawning. [ 1021.679604] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52553e52-0376-47ad-bf07-7958ff5bd169, 'name': SearchDatastore_Task, 'duration_secs': 0.013973} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.680585] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.680585] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.680585] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.680585] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.680823] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.680991] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a018d02-5380-43e3-8471-8f69898117c1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.686517] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.686653] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquired lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.686797] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.690652] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.690921] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1021.691609] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94c1edc4-a8c8-4f97-8f67-4a1318d4e77f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.698072] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1021.698072] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e63388-8897-7d86-5bac-e08dfd6badf3" [ 1021.698072] env[61947]: _type = "Task" [ 1021.698072] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.711440] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e63388-8897-7d86-5bac-e08dfd6badf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.741960] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e0707e-1f96-47ba-9f2a-deb1ef171e86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.749800] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fb0d71-dc1c-40e2-ae44-17dfa05e6a86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.790533] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a32238f-1b6a-496c-9fe9-8b2d5b53b2a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.798547] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c99234-08ad-40e2-8404-37da613c0dab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.811998] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.867894] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.206001} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.868475] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.869267] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7631640-27ee-4c3d-b87f-36c15c063c6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.893487] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 068c9c5e-f7e4-4a5e-896d-a5c348949e07/068c9c5e-f7e4-4a5e-896d-a5c348949e07.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.893831] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43e62a66-0ef1-41aa-9dc2-a2c8b230e45a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.914077] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1021.914077] env[61947]: value = "task-1224871" [ 1021.914077] env[61947]: _type = "Task" [ 1021.914077] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.922152] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224871, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.977229] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.044941] env[61947]: DEBUG nova.compute.manager [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Stashing vm_state: active {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1022.168633] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Updated VIF entry in instance network info cache for port d4f4f476-5bcf-482f-9c37-5436907cd763. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1022.169406] env[61947]: DEBUG nova.network.neutron [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Updating instance_info_cache with network_info: [{"id": "d4f4f476-5bcf-482f-9c37-5436907cd763", "address": "fa:16:3e:96:4c:2c", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4f4f476-5b", "ovs_interfaceid": "d4f4f476-5bcf-482f-9c37-5436907cd763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.209907] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e63388-8897-7d86-5bac-e08dfd6badf3, 'name': SearchDatastore_Task, 'duration_secs': 0.01934} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.210808] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c673dd9f-30b0-42c7-9590-8c8ab6347845 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.216347] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1022.216347] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d21df5-0ac5-8398-417e-df2c5fde4459" [ 1022.216347] env[61947]: _type = "Task" [ 1022.216347] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.224712] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d21df5-0ac5-8398-417e-df2c5fde4459, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.236008] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1022.315577] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.425162] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.442666] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1022.463651] env[61947]: DEBUG nova.network.neutron [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Updating instance_info_cache with network_info: [{"id": "6c189068-6cb8-4da8-8db6-a717706e10fa", "address": "fa:16:3e:ba:09:03", "network": {"id": "8c5d82bd-e6e8-42d2-87cc-99095dcc496b", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-780026279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c07ddc4be8d4411a1762897ecfd89d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c189068-6c", "ovs_interfaceid": "6c189068-6cb8-4da8-8db6-a717706e10fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.474325] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.474596] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.474866] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.474949] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.475158] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.475333] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.475953] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.475953] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.475953] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.476131] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.476349] env[61947]: DEBUG nova.virt.hardware [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.477275] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1696d8-e069-4f26-9efc-78781a557364 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.486036] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af97262f-1fab-492c-8422-837259decd16 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.567796] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.673097] env[61947]: DEBUG oslo_concurrency.lockutils [req-9f4009c0-f8d6-4795-9425-2019e52cc72d req-908120fa-73f8-47f4-9fc8-97ca4272059d service nova] Releasing lock "refresh_cache-a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.726546] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52d21df5-0ac5-8398-417e-df2c5fde4459, 'name': SearchDatastore_Task, 'duration_secs': 0.0114} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.727028] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.727425] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] a56bb67b-93e4-4ec1-a3c6-b9172a4ed617/a56bb67b-93e4-4ec1-a3c6-b9172a4ed617.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1022.728266] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-719c0815-5b24-4675-bb1f-9537f4b3eccb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.735078] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1022.735078] env[61947]: value = "task-1224872" [ 1022.735078] env[61947]: _type = "Task" [ 1022.735078] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.746958] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.822405] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1022.822626] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.335s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.822898] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.255s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.925117] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224871, 'name': ReconfigVM_Task, 'duration_secs': 0.91392} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.925117] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 068c9c5e-f7e4-4a5e-896d-a5c348949e07/068c9c5e-f7e4-4a5e-896d-a5c348949e07.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.925798] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aff9ee13-c127-4774-bdbe-77f731078ef8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.932326] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1022.932326] env[61947]: value = "task-1224873" [ 1022.932326] env[61947]: _type = "Task" [ 1022.932326] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.941720] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224873, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.966982] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Releasing lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.967505] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Instance network_info: |[{"id": "6c189068-6cb8-4da8-8db6-a717706e10fa", "address": "fa:16:3e:ba:09:03", "network": {"id": "8c5d82bd-e6e8-42d2-87cc-99095dcc496b", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-780026279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c07ddc4be8d4411a1762897ecfd89d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c189068-6c", "ovs_interfaceid": "6c189068-6cb8-4da8-8db6-a717706e10fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.968029] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:09:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c189068-6cb8-4da8-8db6-a717706e10fa', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.976570] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Creating folder: Project (7c07ddc4be8d4411a1762897ecfd89d4). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1022.976878] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4c9fb83-7d8f-49e5-8456-faf659b30ba4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.987465] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Created folder: Project (7c07ddc4be8d4411a1762897ecfd89d4) in parent group-v264556. [ 1022.987779] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Creating folder: Instances. Parent ref: group-v264727. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1022.988051] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c0c140c-dedd-4e04-af5c-40421135025d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.997108] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Created folder: Instances in parent group-v264727. [ 1022.997354] env[61947]: DEBUG oslo.service.loopingcall [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.997568] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1022.997791] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-565ceaa2-41ec-42e3-b788-eb9060951875 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.020380] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.020380] env[61947]: value = "task-1224876" [ 1023.020380] env[61947]: _type = "Task" [ 1023.020380] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.031606] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224876, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.045733] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.046159] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.046603] env[61947]: INFO nova.compute.manager [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Attaching volume 73ad2489-f9ec-4a1c-a81a-7b283cab0740 to /dev/sdb [ 1023.076122] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.076496] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.085886] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d85b88-11c7-4e74-89e7-c5b1239f211f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.095401] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b52d4b9-d196-4e43-9362-8cea3b70e71d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.112156] env[61947]: DEBUG nova.virt.block_device [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating existing volume attachment record: 2da1cb8d-4976-44c0-9923-527794d25ace {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 1023.170280] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Successfully updated port: 7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.248719] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224872, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.304087] env[61947]: DEBUG oslo_concurrency.lockutils [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "1660df2a-b2c5-469f-b602-0014dd3656d5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.304391] env[61947]: DEBUG oslo_concurrency.lockutils [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.304588] env[61947]: DEBUG nova.compute.manager [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.305588] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b50c6c-e4f9-4711-b3c4-ae2013be1652 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.314244] env[61947]: DEBUG nova.compute.manager [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1023.314826] env[61947]: DEBUG nova.objects.instance [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'flavor' on Instance uuid 1660df2a-b2c5-469f-b602-0014dd3656d5 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.328284] env[61947]: INFO nova.compute.claims [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.443222] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224873, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.531142] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224876, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.580042] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.580265] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.581215] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000b617f-a48a-43ca-9529-adb3c67f386d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.598650] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66113ea9-30da-49d8-8305-563b1484e3f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.624968] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfiguring VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1023.625436] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee2a529c-5020-4c85-8632-559215f4495c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.643411] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1023.643411] env[61947]: value = "task-1224880" [ 1023.643411] env[61947]: _type = "Task" [ 1023.643411] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.650879] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.673564] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.673710] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.673867] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1023.676983] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Received event network-changed-6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.677173] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Refreshing instance network info cache due to event network-changed-6c189068-6cb8-4da8-8db6-a717706e10fa. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1023.677389] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Acquiring lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.677544] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Acquired lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.677815] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Refreshing network info cache for port 6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1023.748282] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59079} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.748627] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] a56bb67b-93e4-4ec1-a3c6-b9172a4ed617/a56bb67b-93e4-4ec1-a3c6-b9172a4ed617.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1023.748909] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.749189] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-129ba6e6-c29c-45cd-8ad6-8fd91f7a1cf8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.756296] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1023.756296] env[61947]: value = "task-1224881" [ 1023.756296] env[61947]: _type = "Task" [ 1023.756296] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.767030] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.820720] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.821181] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf7aabd7-a893-487a-9ae9-504234ef08d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.828497] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1023.828497] env[61947]: value = "task-1224882" [ 1023.828497] env[61947]: _type = "Task" [ 1023.828497] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.834601] env[61947]: INFO nova.compute.resource_tracker [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating resource usage from migration bdf8ba91-ccf6-4bc1-ab8c-7bc388fd7648 [ 1023.843606] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.944953] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224873, 'name': Rename_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.993119] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb618705-a7cd-4f6b-ab39-a49e1554550c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.000407] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664ef9d3-2e45-4624-92a4-8aecaf07551f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.032308] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9cbd07-b5c8-4e18-a923-1d384290c6b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.039878] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224876, 'name': CreateVM_Task, 'duration_secs': 0.547928} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.041723] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.042458] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.042667] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.042946] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.044130] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d192a1-3914-4723-be4e-758a8adfb68f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.047708] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acfa3ef8-f490-4cd2-8001-8dee2d80b252 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.052367] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1024.052367] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c8686e-fd65-286f-012a-a44005c613e2" [ 1024.052367] env[61947]: _type = "Task" [ 1024.052367] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.059903] env[61947]: DEBUG nova.compute.provider_tree [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.070122] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c8686e-fd65-286f-012a-a44005c613e2, 'name': SearchDatastore_Task, 'duration_secs': 0.00941} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.071010] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.071233] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.071469] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.071620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.071798] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.072301] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d50b307b-f90e-47f6-b49b-aa8537ea5a70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.080473] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.080658] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.081398] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-078ff92c-55c0-473c-959a-99bdc8208005 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.087207] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1024.087207] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523fbd01-c42d-b102-ed21-bc50ab3b1179" [ 1024.087207] env[61947]: _type = "Task" [ 1024.087207] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.097172] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523fbd01-c42d-b102-ed21-bc50ab3b1179, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.153240] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.224789] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1024.265434] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.470498} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.265713] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.266498] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be66aa26-987b-4b98-a146-1bf527b56cbf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.287911] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] a56bb67b-93e4-4ec1-a3c6-b9172a4ed617/a56bb67b-93e4-4ec1-a3c6-b9172a4ed617.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.290250] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-212c5164-e784-4b46-84d0-c4af34a59953 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.310027] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1024.310027] env[61947]: value = "task-1224883" [ 1024.310027] env[61947]: _type = "Task" [ 1024.310027] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.318733] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.338096] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224882, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.452302] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224873, 'name': Rename_Task, 'duration_secs': 1.188201} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.454761] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.455166] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f54d903d-dbaa-4fb9-a68e-eae25856e176 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.462405] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1024.462405] env[61947]: value = "task-1224884" [ 1024.462405] env[61947]: _type = "Task" [ 1024.462405] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.470725] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.509308] env[61947]: DEBUG nova.network.neutron [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updating instance_info_cache with network_info: [{"id": "7af3de27-3309-4b9e-bf96-5615d232914f", "address": "fa:16:3e:e9:5f:1c", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af3de27-33", "ovs_interfaceid": "7af3de27-3309-4b9e-bf96-5615d232914f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.549574] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Updated VIF entry in instance network info cache for port 6c189068-6cb8-4da8-8db6-a717706e10fa. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1024.550072] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Updating instance_info_cache with network_info: [{"id": "6c189068-6cb8-4da8-8db6-a717706e10fa", "address": "fa:16:3e:ba:09:03", "network": {"id": "8c5d82bd-e6e8-42d2-87cc-99095dcc496b", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-780026279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7c07ddc4be8d4411a1762897ecfd89d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c189068-6c", "ovs_interfaceid": "6c189068-6cb8-4da8-8db6-a717706e10fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.566711] env[61947]: DEBUG nova.scheduler.client.report [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.597831] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523fbd01-c42d-b102-ed21-bc50ab3b1179, 'name': SearchDatastore_Task, 'duration_secs': 0.009314} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.598703] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96c1b058-0454-4777-8e16-db3f0b60a9c7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.605230] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1024.605230] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5269441c-e9ee-38cb-354f-7f038e0fa56d" [ 1024.605230] env[61947]: _type = "Task" [ 1024.605230] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.612708] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5269441c-e9ee-38cb-354f-7f038e0fa56d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.655405] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.821936] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224883, 'name': ReconfigVM_Task, 'duration_secs': 0.302933} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.822262] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Reconfigured VM instance instance-00000060 to attach disk [datastore1] a56bb67b-93e4-4ec1-a3c6-b9172a4ed617/a56bb67b-93e4-4ec1-a3c6-b9172a4ed617.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.822855] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-457d92ae-9be0-4c0c-804e-b4d499acab02 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.828810] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1024.828810] env[61947]: value = "task-1224885" [ 1024.828810] env[61947]: _type = "Task" [ 1024.828810] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.839060] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224885, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.841956] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224882, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.972514] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224884, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.014207] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.014551] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Instance network_info: |[{"id": "7af3de27-3309-4b9e-bf96-5615d232914f", "address": "fa:16:3e:e9:5f:1c", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af3de27-33", "ovs_interfaceid": "7af3de27-3309-4b9e-bf96-5615d232914f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.014983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:5f:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7af3de27-3309-4b9e-bf96-5615d232914f', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.022520] env[61947]: DEBUG oslo.service.loopingcall [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.022755] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1025.023013] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a010c8d-d1f8-44aa-81f6-13edb14d755f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.042207] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.042207] env[61947]: value = "task-1224886" [ 1025.042207] env[61947]: _type = "Task" [ 1025.042207] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.049242] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224886, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.052844] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Releasing lock "refresh_cache-d68e1fc3-0129-432b-adce-30773f5bd6ee" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.053085] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Received event network-vif-plugged-7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.053316] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.053577] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.053788] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.053975] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] No waiting events found dispatching network-vif-plugged-7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.054172] env[61947]: WARNING nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Received unexpected event network-vif-plugged-7af3de27-3309-4b9e-bf96-5615d232914f for instance with vm_state building and task_state spawning. [ 1025.054342] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Received event network-changed-7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.054503] env[61947]: DEBUG nova.compute.manager [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Refreshing instance network info cache due to event network-changed-7af3de27-3309-4b9e-bf96-5615d232914f. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1025.054698] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Acquiring lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.054840] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Acquired lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.055024] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Refreshing network info cache for port 7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.072313] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.249s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.072520] env[61947]: INFO nova.compute.manager [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Migrating [ 1025.115534] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5269441c-e9ee-38cb-354f-7f038e0fa56d, 'name': SearchDatastore_Task, 'duration_secs': 0.008172} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.115868] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.116186] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] d68e1fc3-0129-432b-adce-30773f5bd6ee/d68e1fc3-0129-432b-adce-30773f5bd6ee.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1025.116516] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-313754f6-887c-4ee6-9cb0-a1112e75fbd2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.123181] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1025.123181] env[61947]: value = "task-1224887" [ 1025.123181] env[61947]: _type = "Task" [ 1025.123181] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.131152] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.155422] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.345068] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224882, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.348738] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224885, 'name': Rename_Task, 'duration_secs': 0.146954} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.349117] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1025.349420] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-306ed6a1-9599-4c94-a082-9b0a10c3ad81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.355863] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1025.355863] env[61947]: value = "task-1224888" [ 1025.355863] env[61947]: _type = "Task" [ 1025.355863] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.365590] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.474175] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224884, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.556035] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224886, 'name': CreateVM_Task, 'duration_secs': 0.417007} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.556035] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.556257] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.556332] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.556795] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.559864] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afa565af-b25f-4b7d-a369-63f23f36114b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.565585] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1025.565585] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283035c-96bb-5012-cd53-27d0cdbdd3d7" [ 1025.565585] env[61947]: _type = "Task" [ 1025.565585] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.578155] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283035c-96bb-5012-cd53-27d0cdbdd3d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.587007] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.587328] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.587534] env[61947]: DEBUG nova.network.neutron [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.637696] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224887, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.658381] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.772437] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updated VIF entry in instance network info cache for port 7af3de27-3309-4b9e-bf96-5615d232914f. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.772842] env[61947]: DEBUG nova.network.neutron [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updating instance_info_cache with network_info: [{"id": "7af3de27-3309-4b9e-bf96-5615d232914f", "address": "fa:16:3e:e9:5f:1c", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af3de27-33", "ovs_interfaceid": "7af3de27-3309-4b9e-bf96-5615d232914f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.846411] env[61947]: DEBUG oslo_vmware.api [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224882, 'name': PowerOffVM_Task, 'duration_secs': 1.563743} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.846801] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1025.847089] env[61947]: DEBUG nova.compute.manager [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.848230] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a6e218-a7b6-4afc-a291-caf26084c279 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.866386] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224888, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.974766] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224884, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.077064] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5283035c-96bb-5012-cd53-27d0cdbdd3d7, 'name': SearchDatastore_Task, 'duration_secs': 0.043438} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.077377] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.077473] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.077730] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.077887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.078080] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.078345] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8bbe4fdd-914d-4edf-b018-8fb1679f5c66 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.088152] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.088336] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1026.089073] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f10d92f9-479d-4947-9ee3-d3d448202e3e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.096301] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1026.096301] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5202127c-f422-09dc-8ba2-144bf9899adc" [ 1026.096301] env[61947]: _type = "Task" [ 1026.096301] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.104096] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5202127c-f422-09dc-8ba2-144bf9899adc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.132396] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528473} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.133044] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] d68e1fc3-0129-432b-adce-30773f5bd6ee/d68e1fc3-0129-432b-adce-30773f5bd6ee.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1026.133044] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.133231] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77b37ed7-1bab-4b28-ba26-b65e600087fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.139665] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1026.139665] env[61947]: value = "task-1224890" [ 1026.139665] env[61947]: _type = "Task" [ 1026.139665] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.147184] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.154720] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.277816] env[61947]: DEBUG oslo_concurrency.lockutils [req-930cfafe-a74b-4f18-ac4e-cca243fddbba req-5c19179a-577d-4a60-9851-c6c184f298fd service nova] Releasing lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.307133] env[61947]: DEBUG nova.network.neutron [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.364439] env[61947]: DEBUG oslo_concurrency.lockutils [None req-34adb6a7-d8eb-4503-8f6d-ebd6aa43bc8f tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 3.060s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.369023] env[61947]: DEBUG oslo_vmware.api [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224888, 'name': PowerOnVM_Task, 'duration_secs': 0.642098} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.369332] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1026.369537] env[61947]: INFO nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Took 8.55 seconds to spawn the instance on the hypervisor. [ 1026.369766] env[61947]: DEBUG nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.370516] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25a5724-eefb-4890-ba42-7512db18f660 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.475129] env[61947]: DEBUG oslo_vmware.api [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224884, 'name': PowerOnVM_Task, 'duration_secs': 1.543076} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.475410] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1026.475616] env[61947]: INFO nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Took 11.06 seconds to spawn the instance on the hypervisor. [ 1026.475796] env[61947]: DEBUG nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.476572] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a8adc-2592-4afe-a1c3-5dbcd8dae971 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.607961] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5202127c-f422-09dc-8ba2-144bf9899adc, 'name': SearchDatastore_Task, 'duration_secs': 0.020519} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.607961] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a03b69c-56ed-478d-adb4-fbfd48b35c98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.612899] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1026.612899] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254b8bd-208d-6e6f-f3f2-5129b349e53b" [ 1026.612899] env[61947]: _type = "Task" [ 1026.612899] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.620447] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254b8bd-208d-6e6f-f3f2-5129b349e53b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.649120] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.382884} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.652579] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.653325] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527f4827-3b08-49b3-a8fc-bea08f71b19d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.661203] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.680474] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] d68e1fc3-0129-432b-adce-30773f5bd6ee/d68e1fc3-0129-432b-adce-30773f5bd6ee.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.680834] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86f3e545-3f7a-4053-b7c2-133c639c4a80 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.700888] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1026.700888] env[61947]: value = "task-1224891" [ 1026.700888] env[61947]: _type = "Task" [ 1026.700888] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.708908] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224891, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.811105] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.889045] env[61947]: INFO nova.compute.manager [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Took 22.57 seconds to build instance. [ 1026.994708] env[61947]: INFO nova.compute.manager [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Took 24.87 seconds to build instance. [ 1027.004162] env[61947]: DEBUG nova.objects.instance [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'flavor' on Instance uuid 1660df2a-b2c5-469f-b602-0014dd3656d5 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.124955] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5254b8bd-208d-6e6f-f3f2-5129b349e53b, 'name': SearchDatastore_Task, 'duration_secs': 0.046854} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.125280] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.125505] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 1faa141b-ddf6-4fc3-bb83-e77efa022f81/1faa141b-ddf6-4fc3-bb83-e77efa022f81.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1027.126084] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d18ecbf0-b5b7-4770-a693-fa17f9998c6f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.132491] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1027.132491] env[61947]: value = "task-1224892" [ 1027.132491] env[61947]: _type = "Task" [ 1027.132491] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.140616] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224892, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.157799] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.210424] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.391388] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5da8203c-cd52-42b7-b6f5-b7a7a6734ab0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.088s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.495986] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d25651eb-3368-4c7b-a5d2-e1eac909b5ad tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.375s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.509715] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.509888] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquired lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.510086] env[61947]: DEBUG nova.network.neutron [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.510318] env[61947]: DEBUG nova.objects.instance [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'info_cache' on Instance uuid 1660df2a-b2c5-469f-b602-0014dd3656d5 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.521035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.521334] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.521566] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.521757] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.521933] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.524493] env[61947]: INFO nova.compute.manager [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Terminating instance [ 1027.528916] env[61947]: DEBUG nova.compute.manager [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.528916] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.528916] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df118d2f-4aec-4d4e-9920-b1079a8ec557 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.537859] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.538439] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5aa8e7a-ee74-4b18-bb47-0560e8f76d6d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.545667] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1027.545667] env[61947]: value = "task-1224893" [ 1027.545667] env[61947]: _type = "Task" [ 1027.545667] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.557182] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.643500] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224892, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476357} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.643709] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] 1faa141b-ddf6-4fc3-bb83-e77efa022f81/1faa141b-ddf6-4fc3-bb83-e77efa022f81.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.643918] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.644202] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77ac27a1-ab2a-4332-bc29-6f3afd806617 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.654119] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1027.654119] env[61947]: value = "task-1224894" [ 1027.654119] env[61947]: _type = "Task" [ 1027.654119] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.662073] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.667053] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.670323] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1027.670549] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264731', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'name': 'volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9378657d-dc54-47e2-b178-1ed3f9c952f2', 'attached_at': '', 'detached_at': '', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'serial': '73ad2489-f9ec-4a1c-a81a-7b283cab0740'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1027.671359] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74980d59-b803-4857-9941-4462868b8674 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.687080] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de723a8-7cbc-4044-9665-6a33cb99c665 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.713776] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740/volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.717331] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82762756-1b85-4699-a7da-1869b1aea4a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.737455] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224891, 'name': ReconfigVM_Task, 'duration_secs': 0.672279} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.739019] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Reconfigured VM instance instance-00000061 to attach disk [datastore1] d68e1fc3-0129-432b-adce-30773f5bd6ee/d68e1fc3-0129-432b-adce-30773f5bd6ee.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.739801] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1027.739801] env[61947]: value = "task-1224895" [ 1027.739801] env[61947]: _type = "Task" [ 1027.739801] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.740024] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3abae257-47bc-41c1-9cea-286fd3686f24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.751889] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224895, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.753267] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1027.753267] env[61947]: value = "task-1224896" [ 1027.753267] env[61947]: _type = "Task" [ 1027.753267] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.768489] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224896, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.014698] env[61947]: DEBUG nova.objects.base [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Object Instance<1660df2a-b2c5-469f-b602-0014dd3656d5> lazy-loaded attributes: flavor,info_cache {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1028.057565] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224893, 'name': PowerOffVM_Task, 'duration_secs': 0.221633} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.057884] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.058091] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.058351] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-546c7d62-9b7b-41bc-b158-a8b39fbee752 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.149615] env[61947]: DEBUG nova.compute.manager [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Received event network-changed-aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.149964] env[61947]: DEBUG nova.compute.manager [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Refreshing instance network info cache due to event network-changed-aea546b4-1737-4566-8328-b0dcaca4e4e9. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1028.150300] env[61947]: DEBUG oslo_concurrency.lockutils [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] Acquiring lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.150472] env[61947]: DEBUG oslo_concurrency.lockutils [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] Acquired lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.150642] env[61947]: DEBUG nova.network.neutron [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Refreshing network info cache for port aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.166328] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.169858] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067393} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.170524] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.171394] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778c38ca-f464-4b7a-8695-60d3dd82f144 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.196040] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 1faa141b-ddf6-4fc3-bb83-e77efa022f81/1faa141b-ddf6-4fc3-bb83-e77efa022f81.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.196342] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cd8c3fc-061f-48de-ab74-5d8f45f9b71d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.218114] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1028.218114] env[61947]: value = "task-1224898" [ 1028.218114] env[61947]: _type = "Task" [ 1028.218114] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.227550] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224898, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.248037] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.248303] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.248530] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleting the datastore file [datastore1] a56bb67b-93e4-4ec1-a3c6-b9172a4ed617 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.249623] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f301f004-883b-4c02-8aaf-12e9a910a2e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.255164] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.260482] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1028.260482] env[61947]: value = "task-1224899" [ 1028.260482] env[61947]: _type = "Task" [ 1028.260482] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.266879] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224896, 'name': Rename_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.271776] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.327866] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b63dbcc-cc6f-4720-8016-42de448e441d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.349657] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1028.667961] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.728344] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224898, 'name': ReconfigVM_Task, 'duration_secs': 0.316171} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.728662] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 1faa141b-ddf6-4fc3-bb83-e77efa022f81/1faa141b-ddf6-4fc3-bb83-e77efa022f81.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.729328] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-925ee9ed-0942-43a8-9d61-a8820636d989 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.735222] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1028.735222] env[61947]: value = "task-1224900" [ 1028.735222] env[61947]: _type = "Task" [ 1028.735222] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.745459] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224900, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.755720] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224895, 'name': ReconfigVM_Task, 'duration_secs': 0.97027} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.755720] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to attach disk [datastore1] volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740/volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.761216] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4375e7d8-b0a6-4e42-bf0b-578baa4cd79f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.786137] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224896, 'name': Rename_Task, 'duration_secs': 0.727536} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.790107] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1028.792536] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1028.792536] env[61947]: value = "task-1224901" [ 1028.792536] env[61947]: _type = "Task" [ 1028.792536] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.792536] env[61947]: DEBUG oslo_vmware.api [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177823} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.792536] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91ec7ef7-b554-4e61-beeb-cf79b5fc3ab7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.792536] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.792882] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1028.792882] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1028.792967] env[61947]: INFO nova.compute.manager [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1028.793398] env[61947]: DEBUG oslo.service.loopingcall [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.796269] env[61947]: DEBUG nova.compute.manager [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.796364] env[61947]: DEBUG nova.network.neutron [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.803801] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224901, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.805131] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1028.805131] env[61947]: value = "task-1224902" [ 1028.805131] env[61947]: _type = "Task" [ 1028.805131] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.813160] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224902, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.857240] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.857580] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20353f8e-df55-4227-8c41-309d60bc11e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.864029] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1028.864029] env[61947]: value = "task-1224903" [ 1028.864029] env[61947]: _type = "Task" [ 1028.864029] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.872534] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224903, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.090810] env[61947]: DEBUG nova.network.neutron [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Updating instance_info_cache with network_info: [{"id": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "address": "fa:16:3e:40:2c:52", "network": {"id": "d2578cf8-d873-41c8-9636-2f3a37d9f4e3", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-404667450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a821c72e147f7ba7e427113148af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "438671d0-9468-4e44-84c1-4c0ebaa743e0", "external-id": "nsx-vlan-transportzone-918", "segmentation_id": 918, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5dc16b-37", "ovs_interfaceid": "0c5dc16b-3789-41c8-877b-380b3cde8c15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.173127] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.255413] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224900, 'name': Rename_Task, 'duration_secs': 0.150209} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.255883] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.256319] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80d68011-79e7-4142-95d9-347e390acc3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.266987] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1029.266987] env[61947]: value = "task-1224904" [ 1029.266987] env[61947]: _type = "Task" [ 1029.266987] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.279610] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.301801] env[61947]: DEBUG oslo_vmware.api [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224901, 'name': ReconfigVM_Task, 'duration_secs': 0.182993} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.302115] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264731', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'name': 'volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9378657d-dc54-47e2-b178-1ed3f9c952f2', 'attached_at': '', 'detached_at': '', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'serial': '73ad2489-f9ec-4a1c-a81a-7b283cab0740'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1029.318359] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224902, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.374598] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224903, 'name': PowerOffVM_Task, 'duration_secs': 0.247701} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.375542] env[61947]: DEBUG nova.network.neutron [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updated VIF entry in instance network info cache for port aea546b4-1737-4566-8328-b0dcaca4e4e9. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.375883] env[61947]: DEBUG nova.network.neutron [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updating instance_info_cache with network_info: [{"id": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "address": "fa:16:3e:ba:8e:f0", "network": {"id": "247751fc-21e1-4170-b756-d54244d27b23", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1763905409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c530a8353f724243982b78eaa601b131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f0ef5aba-bd9a-42ff-a1a0-5e763986d70a", "external-id": "nsx-vlan-transportzone-209", "segmentation_id": 209, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaea546b4-17", "ovs_interfaceid": "aea546b4-1737-4566-8328-b0dcaca4e4e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.377264] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1029.377494] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1029.594201] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Releasing lock "refresh_cache-1660df2a-b2c5-469f-b602-0014dd3656d5" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.668214] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.777065] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224904, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.816051] env[61947]: DEBUG oslo_vmware.api [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224902, 'name': PowerOnVM_Task, 'duration_secs': 0.580217} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.816051] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.816051] env[61947]: INFO nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Took 9.73 seconds to spawn the instance on the hypervisor. [ 1029.816251] env[61947]: DEBUG nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.816921] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fc3645-25fc-4b07-a52f-d34ec4345484 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.882790] env[61947]: DEBUG oslo_concurrency.lockutils [req-1ab863f9-bbe9-4bd2-86c9-e04c2fd6c3ce req-96eecd08-4134-4a61-b876-a01edc926a89 service nova] Releasing lock "refresh_cache-068c9c5e-f7e4-4a5e-896d-a5c348949e07" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.886157] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.886407] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.886572] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.886756] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.886907] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.887076] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.887287] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.887450] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.887652] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.887839] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.888032] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.893615] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97f2d3d9-daa5-4159-9492-4676d24e394c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.910311] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1029.910311] env[61947]: value = "task-1224905" [ 1029.910311] env[61947]: _type = "Task" [ 1029.910311] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.921749] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.991987] env[61947]: DEBUG nova.network.neutron [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.097723] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.098077] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd4a2fdd-56a9-4890-b11a-22315110041f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.108269] env[61947]: DEBUG oslo_vmware.api [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1030.108269] env[61947]: value = "task-1224906" [ 1030.108269] env[61947]: _type = "Task" [ 1030.108269] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.118485] env[61947]: DEBUG oslo_vmware.api [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.168833] env[61947]: DEBUG oslo_vmware.api [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224880, 'name': ReconfigVM_Task, 'duration_secs': 6.223151} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.169135] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.169391] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Reconfigured VM to detach interface {{(pid=61947) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1030.175011] env[61947]: DEBUG nova.compute.manager [req-7040bc8b-6cae-4375-93bf-244790483348 req-a25e3f78-5767-4048-8e36-acec8c693361 service nova] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Received event network-vif-deleted-d4f4f476-5bcf-482f-9c37-5436907cd763 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.280339] env[61947]: DEBUG oslo_vmware.api [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224904, 'name': PowerOnVM_Task, 'duration_secs': 0.712098} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.280638] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.280897] env[61947]: INFO nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1030.281136] env[61947]: DEBUG nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.281932] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c548a7-86fb-433d-9ea2-e2402a698c47 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.338053] env[61947]: DEBUG nova.objects.instance [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'flavor' on Instance uuid 9378657d-dc54-47e2-b178-1ed3f9c952f2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.342548] env[61947]: INFO nova.compute.manager [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Took 25.92 seconds to build instance. [ 1030.420115] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224905, 'name': ReconfigVM_Task, 'duration_secs': 0.371652} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.420484] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1030.494729] env[61947]: INFO nova.compute.manager [-] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Took 1.70 seconds to deallocate network for instance. [ 1030.621852] env[61947]: DEBUG oslo_vmware.api [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224906, 'name': PowerOnVM_Task, 'duration_secs': 0.506109} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.622258] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.622541] env[61947]: DEBUG nova.compute.manager [None req-3dd9eec9-df82-4192-bd7b-27dd312cad94 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.624065] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b340c6-c471-48a1-a38b-4cf06d3f85ef {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.800653] env[61947]: INFO nova.compute.manager [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Took 17.49 seconds to build instance. [ 1030.843955] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.846056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0dcd69c8-3d49-426d-9200-fce8dcda0157 tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.437s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.846512] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0ef23908-abe2-4c9e-b7ad-41943509e848 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.800s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.847941] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.847941] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.847941] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.848220] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.849646] env[61947]: INFO nova.compute.manager [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Terminating instance [ 1030.851411] env[61947]: DEBUG nova.compute.manager [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1030.851614] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1030.852907] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151fb963-220a-4a04-b2ec-2342e2f8e1a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.860551] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1030.861497] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c1c0bf5-d40f-42c4-a5c7-ed18c8e53649 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.869341] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1030.869341] env[61947]: value = "task-1224907" [ 1030.869341] env[61947]: _type = "Task" [ 1030.869341] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.878713] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.930423] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.930791] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.931171] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.931407] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.931567] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.931726] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.931947] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.932144] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.932699] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.933090] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.933165] env[61947]: DEBUG nova.virt.hardware [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.939309] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1030.940233] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4520de60-07bc-4a85-b223-3e1131e97a00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.959297] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1030.959297] env[61947]: value = "task-1224908" [ 1030.959297] env[61947]: _type = "Task" [ 1030.959297] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.967230] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224908, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.002193] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.002632] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.002976] env[61947]: DEBUG nova.objects.instance [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'resources' on Instance uuid a56bb67b-93e4-4ec1-a3c6-b9172a4ed617 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.302853] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4ca822a7-cd01-45ef-b63c-affce8764eaf tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.006s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.380287] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224907, 'name': PowerOffVM_Task, 'duration_secs': 0.188672} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.380560] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1031.380736] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1031.381014] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2af7202f-2e1f-40e6-a059-8128f811e004 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.443067] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1031.443323] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1031.443507] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Deleting the datastore file [datastore1] d68e1fc3-0129-432b-adce-30773f5bd6ee {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1031.443784] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f42220f-5ed8-46c2-9698-d873900346e3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.451331] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for the task: (returnval){ [ 1031.451331] env[61947]: value = "task-1224910" [ 1031.451331] env[61947]: _type = "Task" [ 1031.451331] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.492329] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.492329] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224908, 'name': ReconfigVM_Task, 'duration_secs': 0.177108} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.492329] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1031.492329] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201ae6d2-7b16-4713-b31f-ee0aa5f28a19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.493933] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.494380] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9704d04f-5ff5-4c48-b58a-ebe3174768a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.518902] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.519295] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.519568] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.520179] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.520391] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.524350] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1031.524350] env[61947]: value = "task-1224911" [ 1031.524350] env[61947]: _type = "Task" [ 1031.524350] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.524836] env[61947]: INFO nova.compute.manager [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Terminating instance [ 1031.530827] env[61947]: DEBUG nova.compute.manager [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.531084] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1031.532131] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d39f9d-1118-44ac-aae9-99132bb0c4fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.545565] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224911, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.545914] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.546170] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25d5dd43-1486-4783-a04b-8b080f473e90 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.552126] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1031.552126] env[61947]: value = "task-1224912" [ 1031.552126] env[61947]: _type = "Task" [ 1031.552126] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.564057] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.597094] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.597434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquired lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.597753] env[61947]: DEBUG nova.network.neutron [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.861261] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d13898-7f6b-49d1-aaf6-6a883a0c76bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.870688] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b7dba-a26f-46ec-abdc-bd3635b1ff00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.901295] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804636a7-996d-4c92-9f92-acbda63dc33a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.909454] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c83b03-b43c-4504-9d17-4c97fc630fb2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.923744] env[61947]: DEBUG nova.compute.provider_tree [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.961392] env[61947]: DEBUG oslo_vmware.api [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Task: {'id': task-1224910, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18625} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.961392] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.961719] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1031.961719] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1031.961947] env[61947]: INFO nova.compute.manager [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1031.962222] env[61947]: DEBUG oslo.service.loopingcall [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.962415] env[61947]: DEBUG nova.compute.manager [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.962511] env[61947]: DEBUG nova.network.neutron [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.036469] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224911, 'name': ReconfigVM_Task, 'duration_secs': 0.313264} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.036747] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627/0da87f27-f431-4ebf-ac5e-571605481627.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.037450] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.065105] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224912, 'name': PowerOffVM_Task, 'duration_secs': 0.2239} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.065105] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.065105] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.065308] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97942f52-4f59-4ddf-ac05-5e9a16697ae6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.139923] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1032.140153] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1032.140470] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleting the datastore file [datastore2] e137f21e-766d-4b20-9d92-5d1907e2baa3 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.140728] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a5fe99c-d844-45ad-8303-60923567e54e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.147640] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1032.147640] env[61947]: value = "task-1224914" [ 1032.147640] env[61947]: _type = "Task" [ 1032.147640] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.160853] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.255791] env[61947]: DEBUG nova.compute.manager [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Stashing vm_state: active {{(pid=61947) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1032.428445] env[61947]: DEBUG nova.scheduler.client.report [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.487557] env[61947]: DEBUG nova.compute.manager [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Received event network-changed-7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.487820] env[61947]: DEBUG nova.compute.manager [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Refreshing instance network info cache due to event network-changed-7af3de27-3309-4b9e-bf96-5615d232914f. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1032.492148] env[61947]: DEBUG oslo_concurrency.lockutils [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] Acquiring lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.492404] env[61947]: DEBUG oslo_concurrency.lockutils [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] Acquired lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.492605] env[61947]: DEBUG nova.network.neutron [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Refreshing network info cache for port 7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.529596] env[61947]: INFO nova.network.neutron [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Port 76c0251c-9621-4df0-83f5-b4a19d1f5da9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1032.530013] env[61947]: DEBUG nova.network.neutron [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [{"id": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "address": "fa:16:3e:ce:53:66", "network": {"id": "720a7469-85cf-4e25-82f3-b574730aae1f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1662632932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb7a249773754feaa2ff8a3b9d5bea54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e4942ef-a8", "ovs_interfaceid": "1e4942ef-a8cf-44ec-bc80-58f9ae5894a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.544661] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb4229b-70c9-45f7-b81b-98e330499f28 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.568608] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1175c9-f115-41ba-b063-2950677e1a33 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.588927] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.661141] env[61947]: DEBUG oslo_vmware.api [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284993} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.661535] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.661814] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1032.662088] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1032.662342] env[61947]: INFO nova.compute.manager [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1032.662774] env[61947]: DEBUG oslo.service.loopingcall [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.662954] env[61947]: DEBUG nova.compute.manager [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.663281] env[61947]: DEBUG nova.network.neutron [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.780022] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.780746] env[61947]: DEBUG nova.network.neutron [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.935936] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.938550] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.161s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.963333] env[61947]: INFO nova.scheduler.client.report [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted allocations for instance a56bb67b-93e4-4ec1-a3c6-b9172a4ed617 [ 1033.034007] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Releasing lock "refresh_cache-e137f21e-766d-4b20-9d92-5d1907e2baa3" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.204391] env[61947]: DEBUG nova.network.neutron [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Port 8e500219-2108-4201-adbc-0378294780ef binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1033.282996] env[61947]: INFO nova.compute.manager [-] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Took 1.32 seconds to deallocate network for instance. [ 1033.431327] env[61947]: DEBUG nova.network.neutron [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updated VIF entry in instance network info cache for port 7af3de27-3309-4b9e-bf96-5615d232914f. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.431924] env[61947]: DEBUG nova.network.neutron [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updating instance_info_cache with network_info: [{"id": "7af3de27-3309-4b9e-bf96-5615d232914f", "address": "fa:16:3e:e9:5f:1c", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af3de27-33", "ovs_interfaceid": "7af3de27-3309-4b9e-bf96-5615d232914f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.445413] env[61947]: INFO nova.compute.claims [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.471626] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7026ae2b-f3f9-4749-843d-ea55d17656a0 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "a56bb67b-93e4-4ec1-a3c6-b9172a4ed617" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.950s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.537599] env[61947]: DEBUG oslo_concurrency.lockutils [None req-74100fdb-926e-432a-b0d5-0f2fc2684f46 tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "interface-e137f21e-766d-4b20-9d92-5d1907e2baa3-76c0251c-9621-4df0-83f5-b4a19d1f5da9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.461s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.789610] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.934940] env[61947]: DEBUG oslo_concurrency.lockutils [req-044e4ec5-d521-4a51-875e-1700d5647a67 req-a8de2959-5247-4f4b-9f14-b56065345cb2 service nova] Releasing lock "refresh_cache-1faa141b-ddf6-4fc3-bb83-e77efa022f81" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.951885] env[61947]: INFO nova.compute.resource_tracker [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating resource usage from migration fc2d2f74-d740-4508-8ed3-57a7a0b8adb4 [ 1034.167445] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a641771a-591b-4f5b-8bb7-c20b63248074 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.176134] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6bbbea-1cee-48cc-891b-8bfe8d78374b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.209283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c541b3-475b-43cf-9bd8-309dc2855034 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.227148] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7c8ac8-3c77-4f67-bbe8-7ece1b78a370 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.236810] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.237102] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.237866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.250114] env[61947]: DEBUG nova.network.neutron [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.251444] env[61947]: DEBUG nova.compute.provider_tree [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.482716] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "4e0a8c79-9445-4474-b516-e697e1854dea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.483111] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.483407] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.483610] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.483781] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.485826] env[61947]: INFO nova.compute.manager [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Terminating instance [ 1034.489138] env[61947]: DEBUG nova.compute.manager [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.489138] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.489909] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3575fc-dc16-4674-b1ac-3dc41666997a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.497753] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1034.498054] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0845f388-85d2-463f-b5e0-19b1b8be7f2a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.505092] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1034.505092] env[61947]: value = "task-1224915" [ 1034.505092] env[61947]: _type = "Task" [ 1034.505092] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.514312] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.521154] env[61947]: DEBUG nova.compute.manager [req-351f2795-214c-486c-9c07-ae72f3cb7915 req-e3743b55-3a52-4d1d-a5ff-54491b970fc3 service nova] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Received event network-vif-deleted-6c189068-6cb8-4da8-8db6-a717706e10fa {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.521404] env[61947]: DEBUG nova.compute.manager [req-351f2795-214c-486c-9c07-ae72f3cb7915 req-e3743b55-3a52-4d1d-a5ff-54491b970fc3 service nova] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Received event network-vif-deleted-1e4942ef-a8cf-44ec-bc80-58f9ae5894a5 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.753556] env[61947]: INFO nova.compute.manager [-] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Took 2.09 seconds to deallocate network for instance. [ 1034.754382] env[61947]: DEBUG nova.scheduler.client.report [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.015410] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224915, 'name': PowerOffVM_Task, 'duration_secs': 0.196365} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.015698] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1035.015874] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.016151] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-142ece92-d686-48e5-92f9-9ac36b236781 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.081078] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1035.081330] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1035.081515] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleting the datastore file [datastore2] 4e0a8c79-9445-4474-b516-e697e1854dea {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.081788] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2291fa8-34b8-4c65-9231-5e866247500a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.088594] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1035.088594] env[61947]: value = "task-1224917" [ 1035.088594] env[61947]: _type = "Task" [ 1035.088594] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.096126] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.263770] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.325s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.264013] env[61947]: INFO nova.compute.manager [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Migrating [ 1035.270948] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.273677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.484s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.273904] env[61947]: DEBUG nova.objects.instance [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lazy-loading 'resources' on Instance uuid d68e1fc3-0129-432b-adce-30773f5bd6ee {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.279833] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.280065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.280265] env[61947]: DEBUG nova.network.neutron [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.598104] env[61947]: DEBUG oslo_vmware.api [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214341} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.598487] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.598562] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.598742] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.598920] env[61947]: INFO nova.compute.manager [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1035.599185] env[61947]: DEBUG oslo.service.loopingcall [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.599383] env[61947]: DEBUG nova.compute.manager [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.599479] env[61947]: DEBUG nova.network.neutron [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.785412] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.785599] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.785773] env[61947]: DEBUG nova.network.neutron [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.004855] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fe5bf2-1bd3-4a44-80b3-abb50de5411d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.013689] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47036e1-f927-4175-947c-05e2c5aff3ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.044320] env[61947]: DEBUG nova.network.neutron [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.046033] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682b54fa-9d82-4cfe-b661-dbf459967834 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.054423] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a36794-70c7-4b3f-baf9-1ff6ed0df65e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.069407] env[61947]: DEBUG nova.compute.provider_tree [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.383354] env[61947]: DEBUG nova.network.neutron [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.550166] env[61947]: DEBUG oslo_concurrency.lockutils [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.572604] env[61947]: DEBUG nova.scheduler.client.report [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.655557] env[61947]: DEBUG nova.compute.manager [req-6bacf53d-618a-458f-9fbb-8d33f2783c83 req-0ab1e390-7469-446f-b90a-6fd244ec76b6 service nova] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Received event network-vif-deleted-3e2fe97e-65e9-40f3-a6f3-4008cd00d83a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.760498] env[61947]: DEBUG nova.network.neutron [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.885958] env[61947]: INFO nova.compute.manager [-] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Took 1.29 seconds to deallocate network for instance. [ 1037.076456] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf50524-d4ae-47d8-a99e-515e2e231080 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.080354] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.082712] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.812s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.082712] env[61947]: DEBUG nova.objects.instance [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'resources' on Instance uuid e137f21e-766d-4b20-9d92-5d1907e2baa3 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.101127] env[61947]: INFO nova.scheduler.client.report [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Deleted allocations for instance d68e1fc3-0129-432b-adce-30773f5bd6ee [ 1037.107692] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0b025a-06ea-4680-a658-fb102884b263 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.115486] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1037.263176] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.393142] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.614264] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e5e42855-3b67-4381-b4dc-90c06c2629cd tempest-InstanceActionsV221TestJSON-1053877477 tempest-InstanceActionsV221TestJSON-1053877477-project-member] Lock "d68e1fc3-0129-432b-adce-30773f5bd6ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.767s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.622500] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1037.622908] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55afa708-f025-4680-a6dc-f38fd41d9c30 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.636269] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1037.636269] env[61947]: value = "task-1224918" [ 1037.636269] env[61947]: _type = "Task" [ 1037.636269] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.652139] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224918, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.783328] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ddb749-dc0e-4112-8783-8408e9693a68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.791199] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9602563e-a813-4789-83cd-ffc37a7e2cf5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.822973] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cbcc9d-e18f-4c43-9143-9de2a3d55efd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.831624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac82e1f-ea54-48eb-97d0-ea5970480f52 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.845573] env[61947]: DEBUG nova.compute.provider_tree [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.146272] env[61947]: DEBUG oslo_vmware.api [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224918, 'name': PowerOnVM_Task, 'duration_secs': 0.383486} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.146587] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.146829] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-27719478-38bd-491b-b23e-5efef9382a72 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance '0da87f27-f431-4ebf-ac5e-571605481627' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1038.350842] env[61947]: DEBUG nova.scheduler.client.report [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.783520] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0841b46c-784f-48a3-8511-323d94773106 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.805229] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 0 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1038.857146] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.859621] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.467s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.859730] env[61947]: DEBUG nova.objects.instance [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'resources' on Instance uuid 4e0a8c79-9445-4474-b516-e697e1854dea {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.882854] env[61947]: INFO nova.scheduler.client.report [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted allocations for instance e137f21e-766d-4b20-9d92-5d1907e2baa3 [ 1039.312321] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1039.312657] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74df9b2d-6883-4f2d-8ab1-4776612d249a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.320277] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1039.320277] env[61947]: value = "task-1224919" [ 1039.320277] env[61947]: _type = "Task" [ 1039.320277] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.328518] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.392239] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8eb0734c-8010-47a7-8225-680e9e643a7f tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "e137f21e-766d-4b20-9d92-5d1907e2baa3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.872s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.548412] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02579f69-c2af-46dd-901c-d769a6fa15ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.556518] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cef28df-7005-4d67-b391-dc6e87dc9567 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.589539] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa37cc9-1554-4410-a9c0-696a2d1bcbbd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.596969] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f266f0-5539-48c7-8e7c-ebc865bf2db8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.610721] env[61947]: DEBUG nova.compute.provider_tree [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.831403] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224919, 'name': PowerOffVM_Task, 'duration_secs': 0.437122} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.831737] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1039.831934] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 17 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1040.113562] env[61947]: DEBUG nova.scheduler.client.report [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.337875] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.338147] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.338292] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.338483] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.338637] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.338787] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.338996] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.339329] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.339528] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.339715] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.339878] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.349714] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4840f44-0669-42a3-a9f6-011dba439754 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.367997] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1040.367997] env[61947]: value = "task-1224920" [ 1040.367997] env[61947]: _type = "Task" [ 1040.367997] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.377272] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224920, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.622795] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.763s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.649034] env[61947]: INFO nova.scheduler.client.report [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted allocations for instance 4e0a8c79-9445-4474-b516-e697e1854dea [ 1040.663859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.663859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.663859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.663859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.663859] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.664987] env[61947]: INFO nova.compute.manager [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Terminating instance [ 1040.667302] env[61947]: DEBUG nova.compute.manager [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.667608] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.668611] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf57457-bdf5-49eb-b7eb-3e650aa0da80 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.677258] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.677764] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bb50d61-4063-4177-874c-6c8ba9b8bef3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.685329] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1040.685329] env[61947]: value = "task-1224921" [ 1040.685329] env[61947]: _type = "Task" [ 1040.685329] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.697521] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.778043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.778043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.778043] env[61947]: DEBUG nova.compute.manager [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Going to confirm migration 5 {{(pid=61947) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1040.879814] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224920, 'name': ReconfigVM_Task, 'duration_secs': 0.434929} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.880200] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 33 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1041.159350] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40cf3360-a35e-4e52-a77c-12c83a18a645 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "4e0a8c79-9445-4474-b516-e697e1854dea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.675s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.197427] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224921, 'name': PowerOffVM_Task, 'duration_secs': 0.217553} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.197427] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1041.197692] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1041.197965] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dace4cbc-45cb-4c68-a9c3-3ac114edd678 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.259373] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1041.259608] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1041.259797] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleting the datastore file [datastore1] fcd2e134-4d5f-43d5-a09c-55a49f8ce48b {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.260125] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a49895d-1966-4aff-94a0-5b29d9d1087e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.267699] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for the task: (returnval){ [ 1041.267699] env[61947]: value = "task-1224923" [ 1041.267699] env[61947]: _type = "Task" [ 1041.267699] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.275590] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.315623] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.315895] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquired lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.316250] env[61947]: DEBUG nova.network.neutron [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.316585] env[61947]: DEBUG nova.objects.instance [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'info_cache' on Instance uuid 0da87f27-f431-4ebf-ac5e-571605481627 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.388210] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.388619] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.388890] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.389244] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.389532] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.389817] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.390210] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.390490] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.390788] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.391079] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.391404] env[61947]: DEBUG nova.virt.hardware [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.400501] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.401483] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97de4452-a935-42ee-941b-92adeab4239e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.429161] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1041.429161] env[61947]: value = "task-1224924" [ 1041.429161] env[61947]: _type = "Task" [ 1041.429161] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.440305] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224924, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.777685] env[61947]: DEBUG oslo_vmware.api [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Task: {'id': task-1224923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122573} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.777956] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.778233] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.778434] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.778615] env[61947]: INFO nova.compute.manager [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1041.778867] env[61947]: DEBUG oslo.service.loopingcall [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.780081] env[61947]: DEBUG nova.compute.manager [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.780162] env[61947]: DEBUG nova.network.neutron [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.938859] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224924, 'name': ReconfigVM_Task, 'duration_secs': 0.189563} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.939156] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1041.939956] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bca8c16-d1c5-40fb-a7d6-99ea75bd1278 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.968205] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.968496] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62529fbb-ce1d-45df-9318-a7cb864d12c6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.988321] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1041.988321] env[61947]: value = "task-1224925" [ 1041.988321] env[61947]: _type = "Task" [ 1041.988321] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.998358] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224925, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.106108] env[61947]: DEBUG nova.compute.manager [req-da64e5a8-30d7-47cb-9d0d-8197baf316b4 req-13038e37-4dfc-4a7f-a3fd-a11deb85acc4 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Received event network-vif-deleted-c32a51a8-436f-4d1d-abb7-452a152efd91 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.106339] env[61947]: INFO nova.compute.manager [req-da64e5a8-30d7-47cb-9d0d-8197baf316b4 req-13038e37-4dfc-4a7f-a3fd-a11deb85acc4 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Neutron deleted interface c32a51a8-436f-4d1d-abb7-452a152efd91; detaching it from the instance and deleting it from the info cache [ 1042.106531] env[61947]: DEBUG nova.network.neutron [req-da64e5a8-30d7-47cb-9d0d-8197baf316b4 req-13038e37-4dfc-4a7f-a3fd-a11deb85acc4 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.349386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "8a4b81d3-d683-4517-806c-88b41ad73c48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.349539] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.349765] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.350688] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.350688] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.354488] env[61947]: INFO nova.compute.manager [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Terminating instance [ 1042.356384] env[61947]: DEBUG nova.compute.manager [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.356584] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1042.357412] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b32eef3-dfed-4abe-a135-5bccdc82ed5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.365137] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1042.366543] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a1278d1-68be-4348-b3ec-26ea99834f5b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.371120] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1042.371120] env[61947]: value = "task-1224926" [ 1042.371120] env[61947]: _type = "Task" [ 1042.371120] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.378670] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.499560] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224925, 'name': ReconfigVM_Task, 'duration_secs': 0.309249} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.499900] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.500271] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 50 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1042.544207] env[61947]: DEBUG nova.network.neutron [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [{"id": "8e500219-2108-4201-adbc-0378294780ef", "address": "fa:16:3e:63:2d:7c", "network": {"id": "8a631413-27c4-4b0d-91f0-0749c33ad29c", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-912316039-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2367a725fa784cc9bbe918bcdf4b12e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e500219-21", "ovs_interfaceid": "8e500219-2108-4201-adbc-0378294780ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.545521] env[61947]: DEBUG nova.network.neutron [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.610197] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f7320369-473f-4f81-a184-0251ce41e23a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.618982] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9ded68-20f8-4b4c-8ed5-d803d85e3c0d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.647076] env[61947]: DEBUG nova.compute.manager [req-da64e5a8-30d7-47cb-9d0d-8197baf316b4 req-13038e37-4dfc-4a7f-a3fd-a11deb85acc4 service nova] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Detach interface failed, port_id=c32a51a8-436f-4d1d-abb7-452a152efd91, reason: Instance fcd2e134-4d5f-43d5-a09c-55a49f8ce48b could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1042.882830] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224926, 'name': PowerOffVM_Task, 'duration_secs': 0.257484} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.882830] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1042.882830] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1042.883185] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2e4bc69-3554-4726-9d46-af0182d8cfa6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.947421] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1042.947755] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1042.947870] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleting the datastore file [datastore2] 8a4b81d3-d683-4517-806c-88b41ad73c48 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.948123] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55f33e3a-95d5-497f-baa2-419d6034ea2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.954034] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1042.954034] env[61947]: value = "task-1224928" [ 1042.954034] env[61947]: _type = "Task" [ 1042.954034] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.961257] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.011911] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dc92e7-1714-4b2b-a45d-82690bfbc645 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.033334] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f0a9e2-a6d0-4069-8ca8-0f87ad9d8429 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.053235] env[61947]: INFO nova.compute.manager [-] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Took 1.27 seconds to deallocate network for instance. [ 1043.053698] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Releasing lock "refresh_cache-0da87f27-f431-4ebf-ac5e-571605481627" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.053983] env[61947]: DEBUG nova.objects.instance [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lazy-loading 'migration_context' on Instance uuid 0da87f27-f431-4ebf-ac5e-571605481627 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.055183] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 67 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1043.464205] env[61947]: DEBUG oslo_vmware.api [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141386} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.464481] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.464670] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1043.464859] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1043.465042] env[61947]: INFO nova.compute.manager [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1043.465307] env[61947]: DEBUG oslo.service.loopingcall [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.465504] env[61947]: DEBUG nova.compute.manager [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.465599] env[61947]: DEBUG nova.network.neutron [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1043.560010] env[61947]: DEBUG nova.objects.base [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Object Instance<0da87f27-f431-4ebf-ac5e-571605481627> lazy-loaded attributes: info_cache,migration_context {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1043.563789] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9850e6e8-eae9-417d-9af2-1685bf051e58 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.567056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.567305] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.567627] env[61947]: DEBUG nova.objects.instance [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lazy-loading 'resources' on Instance uuid fcd2e134-4d5f-43d5-a09c-55a49f8ce48b {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.585608] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-964361f0-2b1b-4ad8-a448-d62716c13313 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.591658] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1043.591658] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527867a8-5c67-b466-8e14-9fecd53c115d" [ 1043.591658] env[61947]: _type = "Task" [ 1043.591658] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.600251] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527867a8-5c67-b466-8e14-9fecd53c115d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.103993] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527867a8-5c67-b466-8e14-9fecd53c115d, 'name': SearchDatastore_Task, 'duration_secs': 0.007883} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.106480] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.147503] env[61947]: DEBUG nova.compute.manager [req-035c420a-165e-4ada-b81d-26bbfe9cc8e3 req-3448260c-178a-4b1e-9f11-2c941a095b4b service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Received event network-vif-deleted-a5b4c8f3-144d-4a20-855b-41e891f9ccda {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.147711] env[61947]: INFO nova.compute.manager [req-035c420a-165e-4ada-b81d-26bbfe9cc8e3 req-3448260c-178a-4b1e-9f11-2c941a095b4b service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Neutron deleted interface a5b4c8f3-144d-4a20-855b-41e891f9ccda; detaching it from the instance and deleting it from the info cache [ 1044.147883] env[61947]: DEBUG nova.network.neutron [req-035c420a-165e-4ada-b81d-26bbfe9cc8e3 req-3448260c-178a-4b1e-9f11-2c941a095b4b service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.208430] env[61947]: DEBUG nova.network.neutron [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.213442] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7f87eb-d59d-4eba-87a4-f09122e3f9f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.222975] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe39a88-b88f-4a1a-8e5e-46c185083dd8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.253041] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ee401e-8597-49f7-a45d-24e4882b9995 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.260761] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2570c639-787e-4bf5-b48a-e7c09bdd1e93 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.273991] env[61947]: DEBUG nova.compute.provider_tree [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.651239] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4e1fc79-69c7-4abb-abf2-6b0ffe4a9d23 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.661061] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64feb2ff-deb4-4811-89bc-29c8984ea67e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.690082] env[61947]: DEBUG nova.compute.manager [req-035c420a-165e-4ada-b81d-26bbfe9cc8e3 req-3448260c-178a-4b1e-9f11-2c941a095b4b service nova] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Detach interface failed, port_id=a5b4c8f3-144d-4a20-855b-41e891f9ccda, reason: Instance 8a4b81d3-d683-4517-806c-88b41ad73c48 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1044.712463] env[61947]: INFO nova.compute.manager [-] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Took 1.25 seconds to deallocate network for instance. [ 1044.729517] env[61947]: DEBUG nova.network.neutron [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1044.777633] env[61947]: DEBUG nova.scheduler.client.report [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.219917] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.282279] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.715s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.285369] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.179s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.304053] env[61947]: INFO nova.scheduler.client.report [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Deleted allocations for instance fcd2e134-4d5f-43d5-a09c-55a49f8ce48b [ 1045.751780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.751780] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.751926] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.811434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-0c33cc29-d849-4bb3-b957-ec84c206f74c tempest-AttachInterfacesTestJSON-212128838 tempest-AttachInterfacesTestJSON-212128838-project-member] Lock "fcd2e134-4d5f-43d5-a09c-55a49f8ce48b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.150s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.911276] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58925301-8691-416c-ab36-65f032879012 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.919359] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e632595-26c4-4b44-9484-55f37b50ea43 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.955429] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3ffd92-8163-4afe-bded-0d815fda2c6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.962828] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc66c39-d44c-4bad-af62-33f836d1698b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.976539] env[61947]: DEBUG nova.compute.provider_tree [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.479393] env[61947]: DEBUG nova.scheduler.client.report [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.719145] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.719425] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.719662] env[61947]: INFO nova.compute.manager [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Shelving [ 1046.789659] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.789659] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.789726] env[61947]: DEBUG nova.network.neutron [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1047.228026] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.228026] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79ddb3c8-7b15-4ab5-9866-1c7bba8fd8e1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.234805] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1047.234805] env[61947]: value = "task-1224929" [ 1047.234805] env[61947]: _type = "Task" [ 1047.234805] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.243733] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.489485] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.204s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.492381] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.273s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.492630] env[61947]: DEBUG nova.objects.instance [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'resources' on Instance uuid 8a4b81d3-d683-4517-806c-88b41ad73c48 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.519954] env[61947]: DEBUG nova.network.neutron [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.745766] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224929, 'name': PowerOffVM_Task, 'duration_secs': 0.198259} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.746045] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.746830] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a37530-1997-404c-8446-76b39284e97e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.766974] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c939a310-b5da-4b69-b2e5-7d3fc942fa44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.024850] env[61947]: DEBUG oslo_concurrency.lockutils [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.058362] env[61947]: INFO nova.scheduler.client.report [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocation for migration bdf8ba91-ccf6-4bc1-ab8c-7bc388fd7648 [ 1048.134608] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274bc060-1493-4a22-bf8e-f04b6ac0fc82 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.142394] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb3c59d-caee-422b-9d55-736f7cb24bff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.171890] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f585324e-71b4-4b79-9f39-e8551595a1c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.180168] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c39f25d-a82f-4711-919a-1843ac6a4da0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.194201] env[61947]: DEBUG nova.compute.provider_tree [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.278022] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1048.278131] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-34bd5c43-7530-42ae-a2e4-484860eccc07 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.287470] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1048.287470] env[61947]: value = "task-1224930" [ 1048.287470] env[61947]: _type = "Task" [ 1048.287470] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.296399] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224930, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.535984] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406e6c11-69bd-46bf-8c7d-bd33fdb723b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.544134] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa2d047-e964-4244-bde1-6413ece24096 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.563581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.786s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.697385] env[61947]: DEBUG nova.scheduler.client.report [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.798631] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224930, 'name': CreateSnapshot_Task, 'duration_secs': 0.410105} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.798862] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1048.799627] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832df56a-374e-4d2d-84e7-fb0e91c13dd6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.202324] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.219887] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.220167] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.220456] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.220670] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.220853] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.223404] env[61947]: INFO nova.scheduler.client.report [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted allocations for instance 8a4b81d3-d683-4517-806c-88b41ad73c48 [ 1049.224806] env[61947]: INFO nova.compute.manager [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Terminating instance [ 1049.229126] env[61947]: DEBUG nova.compute.manager [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.229126] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.229126] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e402d9d0-0152-477d-8a0b-51ecbae644bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.236643] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.236880] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbe5900d-9d0e-4496-a03a-9f4d99341710 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.242729] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1049.242729] env[61947]: value = "task-1224931" [ 1049.242729] env[61947]: _type = "Task" [ 1049.242729] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.250458] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.317179] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1049.318235] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5cdd3647-b35b-42a5-888a-a0f8f8da705d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.327276] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1049.327276] env[61947]: value = "task-1224932" [ 1049.327276] env[61947]: _type = "Task" [ 1049.327276] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.337014] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224932, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.638221] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9198d689-98b2-4b83-bd0c-5d54649adc9b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.660071] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec294aca-7989-4711-9cfd-29bf561076e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.667084] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 83 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1049.735461] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ce74b041-d10a-4ee3-9411-0beaa435a7ef tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "8a4b81d3-d683-4517-806c-88b41ad73c48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.386s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.752834] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224931, 'name': PowerOffVM_Task, 'duration_secs': 0.166076} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.753170] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1049.753311] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1049.753584] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-446b2efd-75bb-4942-b073-940092ad688b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.804786] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.805099] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.805332] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "0da87f27-f431-4ebf-ac5e-571605481627-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.805528] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.805703] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.807747] env[61947]: INFO nova.compute.manager [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Terminating instance [ 1049.809564] env[61947]: DEBUG nova.compute.manager [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1049.809760] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.810680] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fdf9e1-be62-47b3-9abf-9958be35e303 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.818396] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.818657] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e40ef67-96ad-4049-8eba-a8aa2e36573a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.821075] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1049.821274] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1049.821480] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleting the datastore file [datastore1] 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.821996] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01983556-d6a3-4ae7-bc52-5df224dabb55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.826385] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1049.826385] env[61947]: value = "task-1224934" [ 1049.826385] env[61947]: _type = "Task" [ 1049.826385] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.830845] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1049.830845] env[61947]: value = "task-1224935" [ 1049.830845] env[61947]: _type = "Task" [ 1049.830845] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.839411] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.843495] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224932, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.846299] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.174036] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.174441] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8aa73040-1924-40f6-ac69-aea26d4cf246 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.182340] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1050.182340] env[61947]: value = "task-1224936" [ 1050.182340] env[61947]: _type = "Task" [ 1050.182340] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.191221] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.342168] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224934, 'name': PowerOffVM_Task, 'duration_secs': 0.167123} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.345474] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1050.345661] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1050.346254] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224932, 'name': CloneVM_Task} progress is 95%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.346476] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65205707-103c-491f-bc86-690cb20606fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.350599] env[61947]: DEBUG oslo_vmware.api [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224935, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16373} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.351163] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.351419] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1050.351653] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1050.351848] env[61947]: INFO nova.compute.manager [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1050.352096] env[61947]: DEBUG oslo.service.loopingcall [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.352291] env[61947]: DEBUG nova.compute.manager [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.352387] env[61947]: DEBUG nova.network.neutron [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.420080] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1050.420422] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1050.420723] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleting the datastore file [datastore2] 0da87f27-f431-4ebf-ac5e-571605481627 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.422991] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de288932-16c4-40b7-8658-92d8848b4108 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.429454] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for the task: (returnval){ [ 1050.429454] env[61947]: value = "task-1224938" [ 1050.429454] env[61947]: _type = "Task" [ 1050.429454] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.437174] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224938, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.586115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "1660df2a-b2c5-469f-b602-0014dd3656d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.586426] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.586648] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.586839] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.587026] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.589501] env[61947]: INFO nova.compute.manager [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Terminating instance [ 1050.592510] env[61947]: DEBUG nova.compute.manager [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1050.592773] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1050.593648] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c73e4c-c221-496b-a304-34abf865a29e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.600995] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1050.601354] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd232273-9c28-4130-a31a-9cdda957399e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.607100] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1050.607100] env[61947]: value = "task-1224939" [ 1050.607100] env[61947]: _type = "Task" [ 1050.607100] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.615344] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.692113] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.842337] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224932, 'name': CloneVM_Task, 'duration_secs': 1.270981} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.844467] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Created linked-clone VM from snapshot [ 1050.844467] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d01c5c4-20b2-4676-9208-02767dc859ab {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.847883] env[61947]: DEBUG nova.compute.manager [req-35cec51c-250a-4f38-a76a-43901a44be49 req-40fc38aa-70c6-40c4-9724-fe4d647a0c81 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Received event network-vif-deleted-0b7626fb-f420-4ad3-bea5-09a0daafed10 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.848084] env[61947]: INFO nova.compute.manager [req-35cec51c-250a-4f38-a76a-43901a44be49 req-40fc38aa-70c6-40c4-9724-fe4d647a0c81 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Neutron deleted interface 0b7626fb-f420-4ad3-bea5-09a0daafed10; detaching it from the instance and deleting it from the info cache [ 1050.848316] env[61947]: DEBUG nova.network.neutron [req-35cec51c-250a-4f38-a76a-43901a44be49 req-40fc38aa-70c6-40c4-9724-fe4d647a0c81 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.856036] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Uploading image a049b242-eba1-46bb-8b53-de799367d098 {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1050.881384] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1050.881384] env[61947]: value = "vm-264734" [ 1050.881384] env[61947]: _type = "VirtualMachine" [ 1050.881384] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1050.881898] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-428c1e3d-ba10-472a-895a-9a53cedea909 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.889644] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lease: (returnval){ [ 1050.889644] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb8cba-ec95-91a8-7e4e-64f4530ed5ef" [ 1050.889644] env[61947]: _type = "HttpNfcLease" [ 1050.889644] env[61947]: } obtained for exporting VM: (result){ [ 1050.889644] env[61947]: value = "vm-264734" [ 1050.889644] env[61947]: _type = "VirtualMachine" [ 1050.889644] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1050.889931] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the lease: (returnval){ [ 1050.889931] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb8cba-ec95-91a8-7e4e-64f4530ed5ef" [ 1050.889931] env[61947]: _type = "HttpNfcLease" [ 1050.889931] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1050.896323] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1050.896323] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb8cba-ec95-91a8-7e4e-64f4530ed5ef" [ 1050.896323] env[61947]: _type = "HttpNfcLease" [ 1050.896323] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1050.939036] env[61947]: DEBUG oslo_vmware.api [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Task: {'id': task-1224938, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153697} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.939319] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.939508] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1050.939692] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1050.939869] env[61947]: INFO nova.compute.manager [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1050.940133] env[61947]: DEBUG oslo.service.loopingcall [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.940358] env[61947]: DEBUG nova.compute.manager [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.941035] env[61947]: DEBUG nova.network.neutron [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.117101] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224939, 'name': PowerOffVM_Task, 'duration_secs': 0.204619} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.117395] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1051.117571] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1051.117825] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab73c9f6-0ddc-43d4-943c-de9fc9a665dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.176391] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1051.177068] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1051.177326] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleting the datastore file [datastore2] 1660df2a-b2c5-469f-b602-0014dd3656d5 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.177637] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9d2e6fe-a8de-477b-b80d-e4b3b2f00e5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.190147] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for the task: (returnval){ [ 1051.190147] env[61947]: value = "task-1224942" [ 1051.190147] env[61947]: _type = "Task" [ 1051.190147] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.202945] env[61947]: DEBUG oslo_vmware.api [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224936, 'name': PowerOnVM_Task, 'duration_secs': 0.532145} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.203858] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.204129] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d16506-ae40-4ca3-9564-ba4ab52fe8f3 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance '9378657d-dc54-47e2-b178-1ed3f9c952f2' progress to 100 {{(pid=61947) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1051.212029] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.321701] env[61947]: DEBUG nova.network.neutron [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.352220] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4749e02-3906-407f-8ff8-e796f2767019 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.360566] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b2f12f-a500-441e-a951-ba421462ffe0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.386983] env[61947]: DEBUG nova.compute.manager [req-35cec51c-250a-4f38-a76a-43901a44be49 req-40fc38aa-70c6-40c4-9724-fe4d647a0c81 service nova] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Detach interface failed, port_id=0b7626fb-f420-4ad3-bea5-09a0daafed10, reason: Instance 21bb0270-bc20-4ec1-9599-d676845b0dc7 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1051.396244] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1051.396244] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb8cba-ec95-91a8-7e4e-64f4530ed5ef" [ 1051.396244] env[61947]: _type = "HttpNfcLease" [ 1051.396244] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1051.396559] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1051.396559] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bb8cba-ec95-91a8-7e4e-64f4530ed5ef" [ 1051.396559] env[61947]: _type = "HttpNfcLease" [ 1051.396559] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1051.397218] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d1535f-c986-4eb4-ae62-76170e261acc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.404507] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1051.404683] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1051.490089] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9eaa45e7-40d2-45c4-aa53-c59a64db2ab7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.701844] env[61947]: DEBUG oslo_vmware.api [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Task: {'id': task-1224942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133555} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.702323] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.702603] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1051.702983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1051.703285] env[61947]: INFO nova.compute.manager [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1051.703761] env[61947]: DEBUG oslo.service.loopingcall [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.704113] env[61947]: DEBUG nova.network.neutron [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.705323] env[61947]: DEBUG nova.compute.manager [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.705426] env[61947]: DEBUG nova.network.neutron [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.825173] env[61947]: INFO nova.compute.manager [-] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Took 1.47 seconds to deallocate network for instance. [ 1052.208262] env[61947]: INFO nova.compute.manager [-] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Took 1.27 seconds to deallocate network for instance. [ 1052.332030] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.332362] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.332610] env[61947]: DEBUG nova.objects.instance [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'resources' on Instance uuid 21bb0270-bc20-4ec1-9599-d676845b0dc7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.422476] env[61947]: DEBUG nova.network.neutron [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.717178] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.881037] env[61947]: DEBUG nova.compute.manager [req-d221495a-739d-445b-bc26-edba1628d326 req-5bdce026-e1ec-40c1-9f9f-4edfb178ceab service nova] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Received event network-vif-deleted-8e500219-2108-4201-adbc-0378294780ef {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.881582] env[61947]: DEBUG nova.compute.manager [req-d221495a-739d-445b-bc26-edba1628d326 req-5bdce026-e1ec-40c1-9f9f-4edfb178ceab service nova] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Received event network-vif-deleted-0c5dc16b-3789-41c8-877b-380b3cde8c15 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.927129] env[61947]: INFO nova.compute.manager [-] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Took 1.22 seconds to deallocate network for instance. [ 1052.956708] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fee133-a6f4-4cec-9808-db7feff5d373 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.965722] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b1c9d5-fbdc-42b5-9c2a-61be857c9883 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.996951] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d04f39-7234-4d10-9fef-ab35052ea466 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.004882] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fb5933-05ee-4e27-83f9-4c25e98a3edd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.018903] env[61947]: DEBUG nova.compute.provider_tree [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1053.435989] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.541245] env[61947]: ERROR nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [req-fc2f0cf4-432d-45b3-a887-c4cc5c45b0fb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7263fe00-9b30-4433-9e9f-ec265ed2b8f2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fc2f0cf4-432d-45b3-a887-c4cc5c45b0fb"}]} [ 1053.556105] env[61947]: DEBUG nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Refreshing inventories for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1053.570280] env[61947]: DEBUG nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating ProviderTree inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1053.570562] env[61947]: DEBUG nova.compute.provider_tree [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1053.585165] env[61947]: DEBUG nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Refreshing aggregate associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, aggregates: None {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1053.604322] env[61947]: DEBUG nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Refreshing trait associations for resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61947) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1053.721846] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27db1a3b-1841-4673-93ff-f767e28eb0c8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.730153] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba85de6a-11c6-441a-8e1e-2b26079454af {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.762987] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237dfaa5-284a-4326-a60f-090b8f136327 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.773710] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd889779-6b54-44b0-8d3b-ac0872b3e665 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.789469] env[61947]: DEBUG nova.compute.provider_tree [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1053.836851] env[61947]: DEBUG nova.network.neutron [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Port 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 binding to destination host cpu-1 is already ACTIVE {{(pid=61947) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1053.837157] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.837321] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.837491] env[61947]: DEBUG nova.network.neutron [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.319548] env[61947]: DEBUG nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updated inventory for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1054.319848] env[61947]: DEBUG nova.compute.provider_tree [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating resource provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 generation from 133 to 134 during operation: update_inventory {{(pid=61947) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1054.320045] env[61947]: DEBUG nova.compute.provider_tree [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Updating inventory in ProviderTree for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1054.740955] env[61947]: DEBUG nova.network.neutron [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.825645] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.493s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.828056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.111s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.828268] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.830014] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.394s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.830248] env[61947]: DEBUG nova.objects.instance [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lazy-loading 'resources' on Instance uuid 1660df2a-b2c5-469f-b602-0014dd3656d5 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.848474] env[61947]: INFO nova.scheduler.client.report [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Deleted allocations for instance 0da87f27-f431-4ebf-ac5e-571605481627 [ 1054.852427] env[61947]: INFO nova.scheduler.client.report [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleted allocations for instance 21bb0270-bc20-4ec1-9599-d676845b0dc7 [ 1055.245442] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.366382] env[61947]: DEBUG oslo_concurrency.lockutils [None req-4515421a-c0bc-445c-853b-b163860551b9 tempest-DeleteServersTestJSON-1455790156 tempest-DeleteServersTestJSON-1455790156-project-member] Lock "0da87f27-f431-4ebf-ac5e-571605481627" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.561s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.367920] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8dc6e5b3-60ca-491b-bdd3-ec06ffb632cf tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "21bb0270-bc20-4ec1-9599-d676845b0dc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.147s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.427283] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aa080c-d0ab-4132-9b75-fabe2f69ad4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.436303] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449e1b21-dd67-45ec-9509-1a2a6c9b113f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.467255] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1380ad6e-e4c1-4a45-a3d3-119a68f1156c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.475165] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2182b9de-b6b9-4bbf-ae5e-995cc08d6e76 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.487876] env[61947]: DEBUG nova.compute.provider_tree [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.748561] env[61947]: DEBUG nova.compute.manager [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61947) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1055.991204] env[61947]: DEBUG nova.scheduler.client.report [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.496070] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.524606] env[61947]: INFO nova.scheduler.client.report [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Deleted allocations for instance 1660df2a-b2c5-469f-b602-0014dd3656d5 [ 1056.857013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.857413] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.033710] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ea8db8a7-7cce-4868-aded-e0ac56911e49 tempest-ListServerFiltersTestJSON-664637080 tempest-ListServerFiltersTestJSON-664637080-project-member] Lock "1660df2a-b2c5-469f-b602-0014dd3656d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.447s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.360012] env[61947]: DEBUG nova.objects.instance [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'migration_context' on Instance uuid 9378657d-dc54-47e2-b178-1ed3f9c952f2 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.478903] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.482208] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.962861] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5d113a-bb6c-4dff-8bb2-c4be033d2609 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.970845] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa1530a-143d-4077-b51f-5cafcf17229a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.001437] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1058.005506] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73415a2-8ef5-48ac-bba2-0da186cd218a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.013783] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abc47aa-33a5-467e-b817-a2bf06e0dd21 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.027060] env[61947]: DEBUG nova.compute.provider_tree [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.527199] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.530286] env[61947]: DEBUG nova.scheduler.client.report [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.531286] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1059.532414] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602693da-bb90-4e19-91e5-57de9e22b5b9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.538721] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1059.538965] env[61947]: ERROR oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk due to incomplete transfer. [ 1059.539250] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-31632842-57dd-4910-9c50-2362209c2905 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.543687] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.686s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.550058] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.023s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.551876] env[61947]: INFO nova.compute.claims [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.554299] env[61947]: DEBUG oslo_vmware.rw_handles [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5279ec9d-ecec-6079-1e76-44c9e05bd158/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1059.554484] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Uploaded image a049b242-eba1-46bb-8b53-de799367d098 to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1059.556733] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1059.559062] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-24eb8e51-85a4-4d38-956a-4e6e402eb88d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.568608] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1059.568608] env[61947]: value = "task-1224944" [ 1059.568608] env[61947]: _type = "Task" [ 1059.568608] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.576914] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224944, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.079408] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224944, 'name': Destroy_Task, 'duration_secs': 0.322847} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.079709] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Destroyed the VM [ 1060.079867] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1060.080139] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b84bf2fe-9dfd-479d-9220-e89dff5aa835 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.086232] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1060.086232] env[61947]: value = "task-1224945" [ 1060.086232] env[61947]: _type = "Task" [ 1060.086232] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.094685] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224945, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.598963] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224945, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.670726] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cd7d32-d033-4d4a-8041-fd9a8b997052 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.679831] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba34ef1-4f70-4331-9f0c-b6f9f67e66d6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.710130] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b59197a-b413-471f-aea0-43f41859fc88 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.717609] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c472d83f-f07f-4f29-9109-a50d10a2ccd2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.730877] env[61947]: DEBUG nova.compute.provider_tree [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.087008] env[61947]: INFO nova.compute.manager [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Swapping old allocation on dict_keys(['7263fe00-9b30-4433-9e9f-ec265ed2b8f2']) held by migration fc2d2f74-d740-4508-8ed3-57a7a0b8adb4 for instance [ 1061.097729] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224945, 'name': RemoveSnapshot_Task, 'duration_secs': 0.736671} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.099959] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1061.100261] env[61947]: DEBUG nova.compute.manager [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.101214] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57eb222-1c07-4702-a89d-f963133770ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.110024] env[61947]: DEBUG nova.scheduler.client.report [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Overwriting current allocation {'allocations': {'7263fe00-9b30-4433-9e9f-ec265ed2b8f2': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 135}}, 'project_id': '5273b44dfca848538176692ab2a19e3d', 'user_id': '382358439d2c4512b1450ed16bf277bc', 'consumer_generation': 1} on consumer 9378657d-dc54-47e2-b178-1ed3f9c952f2 {{(pid=61947) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1061.194275] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.194509] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.194715] env[61947]: DEBUG nova.network.neutron [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.233560] env[61947]: DEBUG nova.scheduler.client.report [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.617186] env[61947]: INFO nova.compute.manager [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Shelve offloading [ 1061.619470] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1061.619759] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-055fbf3c-1365-49a0-afb1-39c4bbd0dbe3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.626617] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1061.626617] env[61947]: value = "task-1224946" [ 1061.626617] env[61947]: _type = "Task" [ 1061.626617] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.634169] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224946, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.738524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.739138] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.898777] env[61947]: DEBUG nova.network.neutron [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [{"id": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "address": "fa:16:3e:45:12:b6", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ec3f44d-0a", "ovs_interfaceid": "7ec3f44d-0a7d-45d1-8429-6e2eccb59e93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.137045] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1062.137276] env[61947]: DEBUG nova.compute.manager [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.138015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d17023-e852-4bb9-adbd-b3be2f118e5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.143579] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.143774] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.143953] env[61947]: DEBUG nova.network.neutron [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.243815] env[61947]: DEBUG nova.compute.utils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.245210] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1062.245445] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.281667] env[61947]: DEBUG nova.policy [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15d31bf1d0bb46fcabbafbe722265877', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc5cf1feb86a41fcbfb0b1b96d90998e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1062.400771] env[61947]: DEBUG oslo_concurrency.lockutils [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-9378657d-dc54-47e2-b178-1ed3f9c952f2" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.401818] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b957bf3-6419-417f-a241-b16ceef40d99 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.410391] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d2f7a7-5320-4cd0-bbb8-ba09f4f106ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.615677] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Successfully created port: 4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.748343] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.159943] env[61947]: DEBUG nova.network.neutron [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.494975] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1063.495299] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-387e1d07-05f6-4693-80c1-fe21bdd8dc92 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.502904] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1063.502904] env[61947]: value = "task-1224947" [ 1063.502904] env[61947]: _type = "Task" [ 1063.502904] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.510983] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.661547] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.759962] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.783442] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.783698] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.783862] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.784061] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.784216] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.784368] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.784578] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.784744] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.784914] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.785627] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.785870] env[61947]: DEBUG nova.virt.hardware [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.786744] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b172c43-53cf-48e6-b328-25a2b0604ff3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.795961] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d84b052-65bf-4887-9065-de802d5a304f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.927339] env[61947]: DEBUG nova.compute.manager [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-vif-unplugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.927447] env[61947]: DEBUG oslo_concurrency.lockutils [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.927660] env[61947]: DEBUG oslo_concurrency.lockutils [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.927837] env[61947]: DEBUG oslo_concurrency.lockutils [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.928016] env[61947]: DEBUG nova.compute.manager [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] No waiting events found dispatching network-vif-unplugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.928198] env[61947]: WARNING nova.compute.manager [req-3ebe1a31-d943-4ae0-b723-cc29b27536c7 req-3554b8ca-bfb4-434c-8783-157d69bef62e service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received unexpected event network-vif-unplugged-487b8623-6225-453d-9be0-1dd6e1241b97 for instance with vm_state shelved and task_state shelving_offloading. [ 1063.948181] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1063.949057] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8636d0fe-6996-4013-badb-09c565ca1c8f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.956806] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1063.957050] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fded1bc3-adcc-4f61-b3ce-bc0cbda2756d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.012146] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224947, 'name': PowerOffVM_Task, 'duration_secs': 0.292853} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.012407] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1064.013093] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.013306] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.013467] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.013691] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.013872] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.014048] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.014268] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.014431] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.014601] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.014768] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.014943] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.019915] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-291b511a-d84d-4146-b7e9-a1350921d6de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.035009] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1064.035009] env[61947]: value = "task-1224949" [ 1064.035009] env[61947]: _type = "Task" [ 1064.035009] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.047116] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.058504] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1064.058815] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1064.059011] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleting the datastore file [datastore1] 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1064.059280] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c4ffcc2-0efa-400c-840f-6536c378df09 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.066092] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1064.066092] env[61947]: value = "task-1224950" [ 1064.066092] env[61947]: _type = "Task" [ 1064.066092] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.074178] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224950, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.457123] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Successfully updated port: 4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.545086] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224949, 'name': ReconfigVM_Task, 'duration_secs': 0.156952} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.545375] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ed57ea-a5c0-42c9-a349-5c8817b7c3f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.565406] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.565637] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.565802] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.565995] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.566166] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.566320] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.566526] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.566688] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.566858] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.567041] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.567225] env[61947]: DEBUG nova.virt.hardware [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.567997] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-058fcd98-0dcc-4005-8798-741316af799a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.575692] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1064.575692] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527f80ce-439f-5427-a586-385a70b306b4" [ 1064.575692] env[61947]: _type = "Task" [ 1064.575692] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.580134] env[61947]: DEBUG oslo_vmware.api [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224950, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144885} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.583422] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.583627] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1064.583814] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1064.591345] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527f80ce-439f-5427-a586-385a70b306b4, 'name': SearchDatastore_Task, 'duration_secs': 0.009278} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.597158] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1064.597913] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-610b2576-c523-4601-b6e4-b427d7a7555d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.614079] env[61947]: INFO nova.scheduler.client.report [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted allocations for instance 8351ae98-2d33-4f76-a3c1-937738905761 [ 1064.623435] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1064.623435] env[61947]: value = "task-1224951" [ 1064.623435] env[61947]: _type = "Task" [ 1064.623435] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.632173] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.959737] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.960042] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.960110] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.119838] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.119908] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.120117] env[61947]: DEBUG nova.objects.instance [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'resources' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.134689] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224951, 'name': ReconfigVM_Task, 'duration_secs': 0.224225} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.134972] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1065.135837] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2d62b2-a133-43b5-9e91-e6ddf6e4f518 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.162035] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.162364] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55c168ec-da6a-49cb-bf49-ecde310fb4fd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.187076] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1065.187076] env[61947]: value = "task-1224952" [ 1065.187076] env[61947]: _type = "Task" [ 1065.187076] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.197071] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224952, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.507583] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.622670] env[61947]: DEBUG nova.objects.instance [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'numa_topology' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.643028] env[61947]: DEBUG nova.network.neutron [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating instance_info_cache with network_info: [{"id": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "address": "fa:16:3e:a5:85:9f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e33aa48-55", "ovs_interfaceid": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.687968] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.688282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.688514] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.688785] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.688980] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.691070] env[61947]: INFO nova.compute.manager [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Terminating instance [ 1065.695908] env[61947]: DEBUG nova.compute.manager [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1065.696136] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1065.696909] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f102ad-b585-428a-8ffe-67e47e346a68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.704989] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224952, 'name': ReconfigVM_Task, 'duration_secs': 0.287582} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.707015] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2/9378657d-dc54-47e2-b178-1ed3f9c952f2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.707634] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1065.708357] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cba1834-58bd-468c-a40d-a474ca6de826 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.710580] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-386e0024-e925-4f37-b135-e4d388f1d0b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.732571] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c76af1-2a0a-4081-927d-c2fbedb9cf18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.735048] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1065.735048] env[61947]: value = "task-1224953" [ 1065.735048] env[61947]: _type = "Task" [ 1065.735048] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.754733] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e54260f-cf2e-4a7c-a576-fbdd75b916c9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.760070] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.778054] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47586b4a-dedf-418e-94fe-cbbd6e4f951c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.784463] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1065.784714] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a192e0e8-8119-4a37-a83b-891aa4334bc5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.789735] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1065.789735] env[61947]: value = "task-1224954" [ 1065.789735] env[61947]: _type = "Task" [ 1065.789735] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.797179] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.955761] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.955977] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing instance network info cache due to event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1065.956389] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.956553] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.956730] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.125634] env[61947]: DEBUG nova.objects.base [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Object Instance<8351ae98-2d33-4f76-a3c1-937738905761> lazy-loaded attributes: resources,numa_topology {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1066.145675] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.146043] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Instance network_info: |[{"id": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "address": "fa:16:3e:a5:85:9f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e33aa48-55", "ovs_interfaceid": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.146480] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:85:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e33aa48-55f8-4eea-be2f-d3b1f986b08b', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.156890] env[61947]: DEBUG oslo.service.loopingcall [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.159474] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1066.159904] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a454319-5dba-4aaf-bb8a-23b98b5dfd61 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.183235] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.183235] env[61947]: value = "task-1224955" [ 1066.183235] env[61947]: _type = "Task" [ 1066.183235] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.191102] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224955, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.214809] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51610589-34e7-410d-a0a9-69f1d5a93560 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.222320] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4df11c-10c9-4850-9329-372ed28bd74f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.256455] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27cba22-28a9-4d72-a8f0-c4b7eaa8e149 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.266497] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496ee43e-0b61-4bf0-9fc0-09ea8a6fa718 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.269995] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224953, 'name': PowerOffVM_Task, 'duration_secs': 0.182598} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.270268] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1066.270447] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1066.271011] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4248b387-d98c-4e9e-8e47-f53b50a602a3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.280073] env[61947]: DEBUG nova.compute.provider_tree [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.298210] env[61947]: DEBUG oslo_vmware.api [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224954, 'name': PowerOnVM_Task, 'duration_secs': 0.383201} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.298478] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1066.326357] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1066.326591] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1066.326778] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Deleting the datastore file [datastore1] 068c9c5e-f7e4-4a5e-896d-a5c348949e07 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.327073] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a87fe0dd-be4f-4c82-86c5-221939ad0661 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.333240] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for the task: (returnval){ [ 1066.333240] env[61947]: value = "task-1224957" [ 1066.333240] env[61947]: _type = "Task" [ 1066.333240] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.341883] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.345422] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.691710] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updated VIF entry in instance network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.692252] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": null, "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap487b8623-62", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.699510] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224955, 'name': CreateVM_Task, 'duration_secs': 0.447796} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.699656] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1066.700300] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.700470] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.700813] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.701085] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dba112cc-187d-4153-8c54-fa2ce05137fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.706464] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1066.706464] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227f223-0bdd-3cbb-b4b0-a4eb5c590045" [ 1066.706464] env[61947]: _type = "Task" [ 1066.706464] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.714235] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227f223-0bdd-3cbb-b4b0-a4eb5c590045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.783564] env[61947]: DEBUG nova.scheduler.client.report [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.843991] env[61947]: DEBUG oslo_vmware.api [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Task: {'id': task-1224957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141701} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.844292] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.844512] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1066.844753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1066.844991] env[61947]: INFO nova.compute.manager [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1066.845307] env[61947]: DEBUG oslo.service.loopingcall [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.845544] env[61947]: DEBUG nova.compute.manager [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1066.845647] env[61947]: DEBUG nova.network.neutron [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.195185] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.195512] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Received event network-vif-plugged-4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.195689] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.195890] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.196073] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.196249] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] No waiting events found dispatching network-vif-plugged-4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.196420] env[61947]: WARNING nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Received unexpected event network-vif-plugged-4e33aa48-55f8-4eea-be2f-d3b1f986b08b for instance with vm_state building and task_state spawning. [ 1067.196585] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Received event network-changed-4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.196745] env[61947]: DEBUG nova.compute.manager [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Refreshing instance network info cache due to event network-changed-4e33aa48-55f8-4eea-be2f-d3b1f986b08b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1067.196935] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Acquiring lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.197093] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Acquired lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.197259] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Refreshing network info cache for port 4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.217654] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5227f223-0bdd-3cbb-b4b0-a4eb5c590045, 'name': SearchDatastore_Task, 'duration_secs': 0.008568} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.217940] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.218189] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.218423] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.218573] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.218773] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.219048] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6d053a6-7f53-4016-ab87-0816bafed177 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.227551] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.227731] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1067.228449] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f870e32c-bbd2-4eb8-86a2-b60aac5e6b43 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.233941] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1067.233941] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ce4b29-75c7-b5ce-eb21-d26dcbdeaefc" [ 1067.233941] env[61947]: _type = "Task" [ 1067.233941] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.241675] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ce4b29-75c7-b5ce-eb21-d26dcbdeaefc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.289076] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.169s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.343400] env[61947]: INFO nova.compute.manager [None req-934a99dd-51ce-4f41-87e8-65413f69ebe7 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance to original state: 'active' [ 1067.743916] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ce4b29-75c7-b5ce-eb21-d26dcbdeaefc, 'name': SearchDatastore_Task, 'duration_secs': 0.009555} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.747032] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dbd4cff-427d-4a5f-b125-9500cc89d123 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.751483] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1067.751483] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bd003a-5612-74fb-2466-c063246dcb19" [ 1067.751483] env[61947]: _type = "Task" [ 1067.751483] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.758993] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bd003a-5612-74fb-2466-c063246dcb19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.760359] env[61947]: DEBUG nova.network.neutron [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.797885] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d75eeb82-577e-47c9-b666-567a093290d2 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.078s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.798124] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.453s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.798332] env[61947]: INFO nova.compute.manager [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Unshelving [ 1067.951524] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updated VIF entry in instance network info cache for port 4e33aa48-55f8-4eea-be2f-d3b1f986b08b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.951903] env[61947]: DEBUG nova.network.neutron [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating instance_info_cache with network_info: [{"id": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "address": "fa:16:3e:a5:85:9f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e33aa48-55", "ovs_interfaceid": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.978729] env[61947]: DEBUG nova.compute.manager [req-baedb29c-b361-4fa1-8c7e-6c2a0e981490 req-e60b3b7d-18ac-437d-b702-e0b2886905f7 service nova] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Received event network-vif-deleted-aea546b4-1737-4566-8328-b0dcaca4e4e9 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1068.123824] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.125045] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.262628] env[61947]: INFO nova.compute.manager [-] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Took 1.42 seconds to deallocate network for instance. [ 1068.263007] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52bd003a-5612-74fb-2466-c063246dcb19, 'name': SearchDatastore_Task, 'duration_secs': 0.009341} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.264790] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.265075] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] f2bcec82-2730-45bf-b54f-951ba389ae36/f2bcec82-2730-45bf-b54f-951ba389ae36.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1068.268068] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d99a9936-a8ad-45b9-a824-4e40f283a53d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.275054] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1068.275054] env[61947]: value = "task-1224958" [ 1068.275054] env[61947]: _type = "Task" [ 1068.275054] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.284666] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224958, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.455427] env[61947]: DEBUG oslo_concurrency.lockutils [req-3284a69b-220d-42a1-9c40-cb0df8d825ae req-4e3d21cb-2323-48e6-bd78-7591839650c9 service nova] Releasing lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.628079] env[61947]: DEBUG nova.compute.utils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.771903] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.772184] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.772416] env[61947]: DEBUG nova.objects.instance [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lazy-loading 'resources' on Instance uuid 068c9c5e-f7e4-4a5e-896d-a5c348949e07 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.785473] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224958, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443656} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.785714] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] f2bcec82-2730-45bf-b54f-951ba389ae36/f2bcec82-2730-45bf-b54f-951ba389ae36.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1068.785926] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.786183] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fba3cfd1-b038-4a43-a3bb-30881463b9e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.792579] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1068.792579] env[61947]: value = "task-1224959" [ 1068.792579] env[61947]: _type = "Task" [ 1068.792579] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.802026] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224959, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.825056] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.942353] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.942626] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.942855] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.943061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.943288] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.945476] env[61947]: INFO nova.compute.manager [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Terminating instance [ 1068.947452] env[61947]: DEBUG nova.compute.manager [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1068.947669] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1068.947919] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d015669d-8780-4769-b051-d0baa1e142e0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.956050] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1068.956050] env[61947]: value = "task-1224960" [ 1068.956050] env[61947]: _type = "Task" [ 1068.956050] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.964305] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.131201] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.303821] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224959, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067491} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.306191] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.307140] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e873203-19b4-4d8c-93ef-22f936d5eea4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.328538] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] f2bcec82-2730-45bf-b54f-951ba389ae36/f2bcec82-2730-45bf-b54f-951ba389ae36.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.330872] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8726654e-6240-4889-bd65-13c90e752fa6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.350431] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1069.350431] env[61947]: value = "task-1224961" [ 1069.350431] env[61947]: _type = "Task" [ 1069.350431] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.360012] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.374390] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc59771-eb6f-47f2-b4b3-cec03e716126 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.381093] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ee62b8-1d6a-4ea0-b484-2ccd890d7559 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.409858] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eedfede-46fc-4b40-a6bc-c8b811303b19 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.416871] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df96ac12-6ac5-4833-a584-b5234439ad75 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.429472] env[61947]: DEBUG nova.compute.provider_tree [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.463867] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224960, 'name': PowerOffVM_Task, 'duration_secs': 0.228476} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.464146] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1069.464347] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1069.464543] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264731', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'name': 'volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9378657d-dc54-47e2-b178-1ed3f9c952f2', 'attached_at': '2024-10-09T14:53:28.000000', 'detached_at': '', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'serial': '73ad2489-f9ec-4a1c-a81a-7b283cab0740'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1069.465309] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841910cb-0eb2-4f7c-b7d3-c68226ac08a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.486362] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bcc8b4-c50a-4f46-b54c-3aac8c26b0ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.492565] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ed3853-bce9-4e30-b285-816f8e307cad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.512014] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ad8883-0f44-4ca8-ac19-abef5e3eaab1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.526341] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] The volume has not been displaced from its original location: [datastore1] volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740/volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1069.531550] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1069.531855] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dcfbb20-ce7c-468c-ae84-32fc76ff55b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.549220] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1069.549220] env[61947]: value = "task-1224962" [ 1069.549220] env[61947]: _type = "Task" [ 1069.549220] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.556797] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224962, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.860831] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.933161] env[61947]: DEBUG nova.scheduler.client.report [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.058570] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224962, 'name': ReconfigVM_Task, 'duration_secs': 0.216966} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.058740] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1070.063335] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cc96784-6d98-4ba2-83d9-4b5bb7ab912b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.077160] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1070.077160] env[61947]: value = "task-1224963" [ 1070.077160] env[61947]: _type = "Task" [ 1070.077160] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.086529] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.191648] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.191965] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.192242] env[61947]: INFO nova.compute.manager [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Attaching volume 2435a6b4-9dcc-4004-95b3-345df16b8841 to /dev/sdb [ 1070.225500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d7e604-8318-4a45-beec-1e5bdb3d13a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.232414] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d55c750-4398-4aea-99eb-baa37e9726c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.246141] env[61947]: DEBUG nova.virt.block_device [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updating existing volume attachment record: c91e600d-d627-49cb-90bb-d8088a0422c3 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 1070.361025] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224961, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.438194] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.440548] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.616s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.440769] env[61947]: DEBUG nova.objects.instance [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'pci_requests' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.456069] env[61947]: INFO nova.scheduler.client.report [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Deleted allocations for instance 068c9c5e-f7e4-4a5e-896d-a5c348949e07 [ 1070.587052] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224963, 'name': ReconfigVM_Task, 'duration_secs': 0.137733} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.587308] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264731', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'name': 'volume-73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9378657d-dc54-47e2-b178-1ed3f9c952f2', 'attached_at': '2024-10-09T14:53:28.000000', 'detached_at': '', 'volume_id': '73ad2489-f9ec-4a1c-a81a-7b283cab0740', 'serial': '73ad2489-f9ec-4a1c-a81a-7b283cab0740'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1070.587591] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1070.588342] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a062f8ad-380c-46e7-98cd-448a27520c3c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.594454] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.594672] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6afa156-766f-43a6-8bcb-3faf302b68ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.653733] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.653983] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.654223] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleting the datastore file [datastore2] 9378657d-dc54-47e2-b178-1ed3f9c952f2 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.654471] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab58fa6b-d3d2-4514-992c-979a172cd726 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.660948] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1070.660948] env[61947]: value = "task-1224966" [ 1070.660948] env[61947]: _type = "Task" [ 1070.660948] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.668466] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.862149] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224961, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.944293] env[61947]: DEBUG nova.objects.instance [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'numa_topology' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.962730] env[61947]: DEBUG oslo_concurrency.lockutils [None req-33bc4bb3-1adb-4218-8e95-ca107451a9b4 tempest-ServersV294TestFqdnHostnames-1880096130 tempest-ServersV294TestFqdnHostnames-1880096130-project-member] Lock "068c9c5e-f7e4-4a5e-896d-a5c348949e07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.274s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.171406] env[61947]: DEBUG oslo_vmware.api [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14909} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.171711] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.171962] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.172212] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.172403] env[61947]: INFO nova.compute.manager [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1071.172663] env[61947]: DEBUG oslo.service.loopingcall [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.172885] env[61947]: DEBUG nova.compute.manager [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.173029] env[61947]: DEBUG nova.network.neutron [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1071.363346] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224961, 'name': ReconfigVM_Task, 'duration_secs': 1.720187} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.363645] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfigured VM instance instance-00000063 to attach disk [datastore2] f2bcec82-2730-45bf-b54f-951ba389ae36/f2bcec82-2730-45bf-b54f-951ba389ae36.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.364298] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afd58284-8221-4e88-8906-a3daf27d1784 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.371979] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1071.371979] env[61947]: value = "task-1224967" [ 1071.371979] env[61947]: _type = "Task" [ 1071.371979] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.379861] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224967, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.448437] env[61947]: INFO nova.compute.claims [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.883193] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224967, 'name': Rename_Task, 'duration_secs': 0.138429} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.883488] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1071.883740] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe84dccc-e13b-4fb3-a415-e9713250f030 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.890675] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1071.890675] env[61947]: value = "task-1224968" [ 1071.890675] env[61947]: _type = "Task" [ 1071.890675] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.892942] env[61947]: DEBUG nova.compute.manager [req-c4043a3a-8ead-4f77-a5be-fda0cde5b6f8 req-79900245-e0e9-42b0-a25c-2c76eea15579 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Received event network-vif-deleted-7ec3f44d-0a7d-45d1-8429-6e2eccb59e93 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1071.893180] env[61947]: INFO nova.compute.manager [req-c4043a3a-8ead-4f77-a5be-fda0cde5b6f8 req-79900245-e0e9-42b0-a25c-2c76eea15579 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Neutron deleted interface 7ec3f44d-0a7d-45d1-8429-6e2eccb59e93; detaching it from the instance and deleting it from the info cache [ 1071.893326] env[61947]: DEBUG nova.network.neutron [req-c4043a3a-8ead-4f77-a5be-fda0cde5b6f8 req-79900245-e0e9-42b0-a25c-2c76eea15579 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.905019] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.359926] env[61947]: DEBUG nova.network.neutron [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.402133] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224968, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.402380] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0720c8a-de28-4e3f-8e53-e7cd6449a667 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.411015] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6affa9-fa92-4dfa-ba7f-68fe412ebc13 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.434699] env[61947]: DEBUG nova.compute.manager [req-c4043a3a-8ead-4f77-a5be-fda0cde5b6f8 req-79900245-e0e9-42b0-a25c-2c76eea15579 service nova] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Detach interface failed, port_id=7ec3f44d-0a7d-45d1-8429-6e2eccb59e93, reason: Instance 9378657d-dc54-47e2-b178-1ed3f9c952f2 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1072.530696] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b92a22e-f582-4572-8aab-e89fa82239f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.538583] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17353a7d-7a8c-4378-86e5-64044cab3e70 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.569268] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3200ff16-6406-410d-b36a-692522b1c50c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.576551] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1c1df9-7d2f-4e83-83a5-7ade1f19b379 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.590035] env[61947]: DEBUG nova.compute.provider_tree [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.862436] env[61947]: INFO nova.compute.manager [-] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Took 1.69 seconds to deallocate network for instance. [ 1072.902126] env[61947]: DEBUG oslo_vmware.api [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1224968, 'name': PowerOnVM_Task, 'duration_secs': 0.997552} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.902274] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1072.902550] env[61947]: INFO nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Took 9.14 seconds to spawn the instance on the hypervisor. [ 1072.902782] env[61947]: DEBUG nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.903627] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a51907-3134-4859-a371-5048d1cd9469 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.093956] env[61947]: DEBUG nova.scheduler.client.report [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.420717] env[61947]: INFO nova.compute.manager [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Took 14.91 seconds to build instance. [ 1073.469574] env[61947]: INFO nova.compute.manager [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Took 0.61 seconds to detach 1 volumes for instance. [ 1073.598749] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.158s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.638801] env[61947]: INFO nova.network.neutron [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating port 487b8623-6225-453d-9be0-1dd6e1241b97 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1073.922935] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9c83157b-cb9b-495d-b3cc-b538ab338e0f tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.443s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.975692] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.976011] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.976211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.010940] env[61947]: INFO nova.scheduler.client.report [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted allocations for instance 9378657d-dc54-47e2-b178-1ed3f9c952f2 [ 1074.273062] env[61947]: DEBUG nova.compute.manager [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Received event network-changed-4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.273279] env[61947]: DEBUG nova.compute.manager [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Refreshing instance network info cache due to event network-changed-4e33aa48-55f8-4eea-be2f-d3b1f986b08b. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1074.273517] env[61947]: DEBUG oslo_concurrency.lockutils [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] Acquiring lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.273666] env[61947]: DEBUG oslo_concurrency.lockutils [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] Acquired lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.273830] env[61947]: DEBUG nova.network.neutron [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Refreshing network info cache for port 4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1074.521557] env[61947]: DEBUG oslo_concurrency.lockutils [None req-650e7afd-04be-4e01-943e-6f220cda99be tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "9378657d-dc54-47e2-b178-1ed3f9c952f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.579s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.791022] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1074.791302] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264736', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'name': 'volume-2435a6b4-9dcc-4004-95b3-345df16b8841', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1faa141b-ddf6-4fc3-bb83-e77efa022f81', 'attached_at': '', 'detached_at': '', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'serial': '2435a6b4-9dcc-4004-95b3-345df16b8841'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1074.792570] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4350de0f-adc9-4323-bc95-8e37fe9bd39b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.818462] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887a83af-538e-4362-9097-929c1ae50d1a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.847117] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-2435a6b4-9dcc-4004-95b3-345df16b8841/volume-2435a6b4-9dcc-4004-95b3-345df16b8841.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.847469] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-499ea84d-6958-49ab-bb43-996e7b55d79f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.867009] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1074.867009] env[61947]: value = "task-1224970" [ 1074.867009] env[61947]: _type = "Task" [ 1074.867009] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.876423] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.333032] env[61947]: DEBUG nova.compute.manager [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1075.333275] env[61947]: DEBUG oslo_concurrency.lockutils [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.335039] env[61947]: DEBUG oslo_concurrency.lockutils [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.335039] env[61947]: DEBUG oslo_concurrency.lockutils [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.335039] env[61947]: DEBUG nova.compute.manager [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] No waiting events found dispatching network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1075.335039] env[61947]: WARNING nova.compute.manager [req-440b4ea4-77fd-44fd-91da-dc5dd0357086 req-63713971-8121-47e5-a7b4-5417a8c296a2 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received unexpected event network-vif-plugged-487b8623-6225-453d-9be0-1dd6e1241b97 for instance with vm_state shelved_offloaded and task_state spawning. [ 1075.352739] env[61947]: DEBUG nova.network.neutron [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updated VIF entry in instance network info cache for port 4e33aa48-55f8-4eea-be2f-d3b1f986b08b. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1075.353312] env[61947]: DEBUG nova.network.neutron [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating instance_info_cache with network_info: [{"id": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "address": "fa:16:3e:a5:85:9f", "network": {"id": "53663ef9-35bf-4225-9a94-5af45b56b5f2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1771083694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc5cf1feb86a41fcbfb0b1b96d90998e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e33aa48-55", "ovs_interfaceid": "4e33aa48-55f8-4eea-be2f-d3b1f986b08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.387305] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224970, 'name': ReconfigVM_Task, 'duration_secs': 0.409473} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.387305] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-2435a6b4-9dcc-4004-95b3-345df16b8841/volume-2435a6b4-9dcc-4004-95b3-345df16b8841.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.394174] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfcc8ca1-919d-4b9e-a11d-69abef93351f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.414625] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1075.414625] env[61947]: value = "task-1224971" [ 1075.414625] env[61947]: _type = "Task" [ 1075.414625] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.423770] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224971, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.455567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.455567] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.455567] env[61947]: DEBUG nova.network.neutron [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.860049] env[61947]: DEBUG oslo_concurrency.lockutils [req-73e5bc62-ba49-4448-a0ed-813414fdc43b req-6f762005-5040-48ea-913a-4b495df85539 service nova] Releasing lock "refresh_cache-f2bcec82-2730-45bf-b54f-951ba389ae36" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.925032] env[61947]: DEBUG oslo_vmware.api [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224971, 'name': ReconfigVM_Task, 'duration_secs': 0.133749} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.925239] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264736', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'name': 'volume-2435a6b4-9dcc-4004-95b3-345df16b8841', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1faa141b-ddf6-4fc3-bb83-e77efa022f81', 'attached_at': '', 'detached_at': '', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'serial': '2435a6b4-9dcc-4004-95b3-345df16b8841'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1076.188593] env[61947]: DEBUG nova.network.neutron [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.323685] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.323876] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.673591] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.673854] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.691614] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.716879] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='625eb18e4a2e6b3474b73a642217504e',container_format='bare',created_at=2024-10-09T14:53:12Z,direct_url=,disk_format='vmdk',id=a049b242-eba1-46bb-8b53-de799367d098,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1757515780-shelved',owner='1d76ebc8875f4b03bc768320b44b11d1',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-09T14:53:25Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.717151] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.717318] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.717509] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.717658] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.717806] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.718085] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.718299] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.718480] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.718647] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.718822] env[61947]: DEBUG nova.virt.hardware [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.719906] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a7b27b-02a3-41f0-a78c-c3b4683b35f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.727981] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e494e6e2-a0ff-40b1-b397-f3d5224b4404 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.741199] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:65:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '487b8623-6225-453d-9be0-1dd6e1241b97', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.748467] env[61947]: DEBUG oslo.service.loopingcall [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.748701] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1076.748899] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f698cbf-3375-4621-958d-e86e46d3b011 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.767804] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.767804] env[61947]: value = "task-1224972" [ 1076.767804] env[61947]: _type = "Task" [ 1076.767804] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.777110] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224972, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.826438] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.968490] env[61947]: DEBUG nova.objects.instance [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid 1faa141b-ddf6-4fc3-bb83-e77efa022f81 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.176400] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.277932] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224972, 'name': CreateVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.352142] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.352436] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.354008] env[61947]: INFO nova.compute.claims [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.360697] env[61947]: DEBUG nova.compute.manager [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.360900] env[61947]: DEBUG nova.compute.manager [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing instance network info cache due to event network-changed-487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.361347] env[61947]: DEBUG oslo_concurrency.lockutils [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.361473] env[61947]: DEBUG oslo_concurrency.lockutils [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.361677] env[61947]: DEBUG nova.network.neutron [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Refreshing network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.474010] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1bb8b149-a07c-464c-80d7-68b110be9c9c tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.282s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.697648] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.778862] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224972, 'name': CreateVM_Task, 'duration_secs': 0.567448} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.780054] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1077.780054] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.780054] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.780446] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.780693] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a55fd7cc-af30-4fe9-a597-f5bdd60ff30c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.784872] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1077.784872] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5213d3f6-878b-9f78-528b-0260cb383716" [ 1077.784872] env[61947]: _type = "Task" [ 1077.784872] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.792209] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5213d3f6-878b-9f78-528b-0260cb383716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.832740] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.832962] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.059828] env[61947]: DEBUG nova.network.neutron [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updated VIF entry in instance network info cache for port 487b8623-6225-453d-9be0-1dd6e1241b97. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1078.060222] env[61947]: DEBUG nova.network.neutron [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.295417] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.295677] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Processing image a049b242-eba1-46bb-8b53-de799367d098 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.295915] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.296112] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.296300] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.296550] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7eeffa8-8a74-4fa9-a081-776525f3131e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.304421] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.305030] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1078.305374] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35091839-2da1-4a91-99b6-c684c5b09ac0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.311089] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1078.311089] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a00ec7-ccd5-4886-c0df-5a3295bfc84f" [ 1078.311089] env[61947]: _type = "Task" [ 1078.311089] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.318148] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a00ec7-ccd5-4886-c0df-5a3295bfc84f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.336382] env[61947]: INFO nova.compute.manager [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Detaching volume 2435a6b4-9dcc-4004-95b3-345df16b8841 [ 1078.369828] env[61947]: INFO nova.virt.block_device [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Attempting to driver detach volume 2435a6b4-9dcc-4004-95b3-345df16b8841 from mountpoint /dev/sdb [ 1078.370085] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1078.370283] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264736', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'name': 'volume-2435a6b4-9dcc-4004-95b3-345df16b8841', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1faa141b-ddf6-4fc3-bb83-e77efa022f81', 'attached_at': '', 'detached_at': '', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'serial': '2435a6b4-9dcc-4004-95b3-345df16b8841'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1078.371214] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c1abba-6d06-4824-b651-e1a27b3085c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.393832] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348255fb-dd71-4e1d-9fc0-e27767ffbb7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.400610] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2ba684-4243-4bc4-9a94-a328f92847e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.426069] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e4dd27-cf7c-4279-bc7f-f17837637ea5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.440730] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] The volume has not been displaced from its original location: [datastore2] volume-2435a6b4-9dcc-4004-95b3-345df16b8841/volume-2435a6b4-9dcc-4004-95b3-345df16b8841.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1078.445935] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1078.448371] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da1d61e8-d217-4db2-8ffb-a77f6b7d2901 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.466400] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1078.466400] env[61947]: value = "task-1224973" [ 1078.466400] env[61947]: _type = "Task" [ 1078.466400] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.476289] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224973, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.482867] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406d37d6-65b6-46bf-8d5c-d4e0a693c9ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.491276] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2032a60-4fb7-4858-91c3-23e590049da8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.522122] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52f1c00-aa1d-449c-aed0-68bcb3e7f203 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.529327] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6106b4-6814-4d21-a477-b41e25606937 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.542749] env[61947]: DEBUG nova.compute.provider_tree [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.563387] env[61947]: DEBUG oslo_concurrency.lockutils [req-8ff31158-63c4-439b-9234-33ba1f82ff12 req-649e0bce-690c-4a2d-bc50-08c5a5fea4f0 service nova] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.820756] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1078.821048] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Fetch image to [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c/OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1078.821248] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Downloading stream optimized image a049b242-eba1-46bb-8b53-de799367d098 to [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c/OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c.vmdk on the data store datastore2 as vApp {{(pid=61947) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1078.821426] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Downloading image file data a049b242-eba1-46bb-8b53-de799367d098 to the ESX as VM named 'OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c' {{(pid=61947) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1078.893585] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1078.893585] env[61947]: value = "resgroup-9" [ 1078.893585] env[61947]: _type = "ResourcePool" [ 1078.893585] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1078.893895] env[61947]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5b7ba1e0-5690-45aa-9128-70fc2ac8846d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.915233] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lease: (returnval){ [ 1078.915233] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1078.915233] env[61947]: _type = "HttpNfcLease" [ 1078.915233] env[61947]: } obtained for vApp import into resource pool (val){ [ 1078.915233] env[61947]: value = "resgroup-9" [ 1078.915233] env[61947]: _type = "ResourcePool" [ 1078.915233] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1078.915492] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the lease: (returnval){ [ 1078.915492] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1078.915492] env[61947]: _type = "HttpNfcLease" [ 1078.915492] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1078.921302] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1078.921302] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1078.921302] env[61947]: _type = "HttpNfcLease" [ 1078.921302] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1078.977157] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224973, 'name': ReconfigVM_Task, 'duration_secs': 0.225375} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.977496] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1078.982515] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3652a03-94b6-4ece-bebc-4e6f4717add9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.997586] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1078.997586] env[61947]: value = "task-1224975" [ 1078.997586] env[61947]: _type = "Task" [ 1078.997586] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.005539] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224975, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.045739] env[61947]: DEBUG nova.scheduler.client.report [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.424447] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1079.424447] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1079.424447] env[61947]: _type = "HttpNfcLease" [ 1079.424447] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1079.507829] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224975, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.549960] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.550521] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.553417] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.856s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.554887] env[61947]: INFO nova.compute.claims [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1079.924906] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1079.924906] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1079.924906] env[61947]: _type = "HttpNfcLease" [ 1079.924906] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1079.925531] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1079.925531] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5253d52d-57d3-6fee-f01b-2dfc9a1d903f" [ 1079.925531] env[61947]: _type = "HttpNfcLease" [ 1079.925531] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1079.926350] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b3e933-a933-4adc-98ed-1c6907b2e756 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.933613] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1079.933802] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1079.998129] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-158d2427-7b16-40c9-b9a1-6d2728223212 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.009337] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224975, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.059675] env[61947]: DEBUG nova.compute.utils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.063055] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.063232] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1080.104248] env[61947]: DEBUG nova.policy [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '382358439d2c4512b1450ed16bf277bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5273b44dfca848538176692ab2a19e3d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1080.419985] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Successfully created port: 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.510311] env[61947]: DEBUG oslo_vmware.api [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224975, 'name': ReconfigVM_Task, 'duration_secs': 1.275591} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.510636] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264736', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'name': 'volume-2435a6b4-9dcc-4004-95b3-345df16b8841', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1faa141b-ddf6-4fc3-bb83-e77efa022f81', 'attached_at': '', 'detached_at': '', 'volume_id': '2435a6b4-9dcc-4004-95b3-345df16b8841', 'serial': '2435a6b4-9dcc-4004-95b3-345df16b8841'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1080.563773] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.656897] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722b2608-b593-4842-a37a-275f4f9dd018 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.666938] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d1a3ce-9d11-4b26-8a67-e08885123a77 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.706614] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c415db6a-4714-4ec1-8438-1dbd85feb8ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.714852] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064fbc5b-884c-4c4d-9131-50c0de961b22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.730928] env[61947]: DEBUG nova.compute.provider_tree [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.061782] env[61947]: DEBUG nova.objects.instance [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid 1faa141b-ddf6-4fc3-bb83-e77efa022f81 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.087522] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1081.087742] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1081.088649] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dc5866-3786-4729-b8fa-7781e5fcd8bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.095420] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1081.095598] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1081.095879] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-892162ba-b93e-4565-af09-879f509d83e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.236235] env[61947]: DEBUG nova.scheduler.client.report [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.271808] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d18d4-fabc-8d99-799a-b9f5cc5e8bce/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1081.272055] env[61947]: INFO nova.virt.vmwareapi.images [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Downloaded image file data a049b242-eba1-46bb-8b53-de799367d098 [ 1081.272861] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b9b40d-2b9d-43eb-9251-ef1396e8ab97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.287993] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba502afc-ddd2-4e07-80cd-a773035e4537 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.316062] env[61947]: INFO nova.virt.vmwareapi.images [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] The imported VM was unregistered [ 1081.318524] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1081.318767] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Creating directory with path [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.319050] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb695d09-0cb5-47bf-8ae2-e11508f32e01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.330421] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Created directory with path [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.330613] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c/OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c.vmdk to [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk. {{(pid=61947) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1081.330864] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f746fabd-5aac-4010-8e0e-4e29cd32576d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.337942] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1081.337942] env[61947]: value = "task-1224977" [ 1081.337942] env[61947]: _type = "Task" [ 1081.337942] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.345822] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.576033] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.604635] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.604995] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.605238] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.605463] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.605646] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.605849] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.606117] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.606335] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.606538] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.606751] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.606977] env[61947]: DEBUG nova.virt.hardware [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.608040] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb2ac74-5436-45b0-beeb-bcd67dcffe7a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.618034] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9906617e-315c-4819-8e55-e97cef5e08db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.741793] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.742598] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1081.845410] env[61947]: DEBUG nova.compute.manager [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1081.845658] env[61947]: DEBUG oslo_concurrency.lockutils [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.845940] env[61947]: DEBUG oslo_concurrency.lockutils [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.846271] env[61947]: DEBUG oslo_concurrency.lockutils [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.846513] env[61947]: DEBUG nova.compute.manager [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] No waiting events found dispatching network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.846709] env[61947]: WARNING nova.compute.manager [req-37f9d189-1c77-4af2-abe7-d86ea8d042a7 req-0f8682b8-017e-4b0d-a16d-7225e24fdaaf service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received unexpected event network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 for instance with vm_state building and task_state spawning. [ 1081.852738] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.944774] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Successfully updated port: 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.071270] env[61947]: DEBUG oslo_concurrency.lockutils [None req-fb577e8e-0519-449d-8a36-8527ad1dfde2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.238s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.248072] env[61947]: DEBUG nova.compute.utils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.249531] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1082.249675] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1082.317402] env[61947]: DEBUG nova.policy [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faf8b6b46e514417a4cf4d82aa21de42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1469b7f9a7f4d74ba0580b72c967feb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1082.348986] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.446786] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.447075] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.447169] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.687700] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Successfully created port: 2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1082.754185] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1082.825388] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.825608] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.825976] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1082.851205] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.986085] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1083.125802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.126110] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.126329] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.126519] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.126694] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.129358] env[61947]: INFO nova.compute.manager [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Terminating instance [ 1083.131349] env[61947]: DEBUG nova.compute.manager [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1083.131548] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1083.132500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f47d228-64a5-4bc3-b3af-ed3a04315f62 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.140472] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.140695] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7da8d64-1fb6-4a86-bf95-54ed9609c08b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.148296] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1083.148296] env[61947]: value = "task-1224978" [ 1083.148296] env[61947]: _type = "Task" [ 1083.148296] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.153057] env[61947]: DEBUG nova.network.neutron [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.160027] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.350585] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.658074] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.658393] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance network_info: |[{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.659065] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224978, 'name': PowerOffVM_Task, 'duration_secs': 0.23447} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.659461] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:43:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c342e49-9cb5-4596-ae93-6dd705d98225', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.667038] env[61947]: DEBUG oslo.service.loopingcall [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.667263] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.667436] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1083.667681] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1083.667895] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d943b2f-1eb2-4c8a-9f20-3c86761d1d36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.669556] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e40e914a-7519-4700-9db1-04ce898ccfc4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.689584] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.689584] env[61947]: value = "task-1224980" [ 1083.689584] env[61947]: _type = "Task" [ 1083.689584] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.697969] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224980, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.742547] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1083.742839] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1083.743079] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleting the datastore file [datastore1] 1faa141b-ddf6-4fc3-bb83-e77efa022f81 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.743389] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11bfcb24-b8e4-41e9-b5d8-7f7fa8d26cdb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.749554] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1083.749554] env[61947]: value = "task-1224981" [ 1083.749554] env[61947]: _type = "Task" [ 1083.749554] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.757749] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.763064] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1083.787863] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1083.788140] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1083.788358] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1083.788564] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1083.788822] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1083.789036] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1083.789307] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1083.789496] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1083.790045] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1083.790045] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1083.790045] env[61947]: DEBUG nova.virt.hardware [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.790866] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d73222-64b4-4b08-a928-73b28ac3ad06 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.798764] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89120fb7-5dbe-454d-8bfe-cef6db31457c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.851696] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224977, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.340047} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.852017] env[61947]: INFO nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c/OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c.vmdk to [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk. [ 1083.852264] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Cleaning up location [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1083.852480] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a3d727eb-6c2f-48b6-9203-e57d5aa4456c {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.852799] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-773d7c36-0d60-4163-8e4f-5ea67bc3a8ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.859992] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1083.859992] env[61947]: value = "task-1224982" [ 1083.859992] env[61947]: _type = "Task" [ 1083.859992] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.870297] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224982, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.872521] env[61947]: DEBUG nova.compute.manager [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1083.872720] env[61947]: DEBUG nova.compute.manager [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing instance network info cache due to event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1083.872991] env[61947]: DEBUG oslo_concurrency.lockutils [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.873213] env[61947]: DEBUG oslo_concurrency.lockutils [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.873456] env[61947]: DEBUG nova.network.neutron [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.199668] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224980, 'name': CreateVM_Task, 'duration_secs': 0.365915} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.200521] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Successfully updated port: 2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.201601] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1084.205021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.205021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.205021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.205021] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732313ab-4370-4853-98a7-08c174c34207 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.208238] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1084.208238] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52728026-858f-ccb2-766a-97d1482331fa" [ 1084.208238] env[61947]: _type = "Task" [ 1084.208238] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.217634] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52728026-858f-ccb2-766a-97d1482331fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.258514] env[61947]: DEBUG oslo_vmware.api [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1224981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143862} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.258767] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.258952] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1084.259211] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1084.259405] env[61947]: INFO nova.compute.manager [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1084.259651] env[61947]: DEBUG oslo.service.loopingcall [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.259848] env[61947]: DEBUG nova.compute.manager [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1084.259945] env[61947]: DEBUG nova.network.neutron [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1084.370019] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224982, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037457} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.370349] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.370529] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.370779] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk to [datastore2] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1084.371053] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6c33fd7-942c-4aa0-aeb2-01ac3e076507 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.378724] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1084.378724] env[61947]: value = "task-1224983" [ 1084.378724] env[61947]: _type = "Task" [ 1084.378724] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.386666] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.705037] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.705037] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquired lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.705037] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.718987] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52728026-858f-ccb2-766a-97d1482331fa, 'name': SearchDatastore_Task, 'duration_secs': 0.026435} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.719453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.719738] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.720013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.720197] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.720434] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.720779] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b881fa77-141f-4639-b000-571cdcaa5de2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.729435] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.729647] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1084.730386] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71bce86d-d465-45a4-a204-b471e5d42488 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.735838] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1084.735838] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ae9f09-8151-b457-c52d-70768ff49053" [ 1084.735838] env[61947]: _type = "Task" [ 1084.735838] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.743533] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ae9f09-8151-b457-c52d-70768ff49053, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.747062] env[61947]: DEBUG nova.network.neutron [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updated VIF entry in instance network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.747062] env[61947]: DEBUG nova.network.neutron [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.837047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.837224] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.839125] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1084.888569] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.240936] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1085.249357] env[61947]: DEBUG oslo_concurrency.lockutils [req-180dc54f-0875-4b2f-8da7-7171b87c1921 req-c20eadc9-46bf-4cf1-9039-101156ccb368 service nova] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.249814] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ae9f09-8151-b457-c52d-70768ff49053, 'name': SearchDatastore_Task, 'duration_secs': 0.01826} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.250642] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00c0ff15-53a6-4854-a72f-17d6d48cda81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.256163] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1085.256163] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282a49e-248c-6aae-d5d6-31ca5bf610d4" [ 1085.256163] env[61947]: _type = "Task" [ 1085.256163] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.266161] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282a49e-248c-6aae-d5d6-31ca5bf610d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.390132] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.414387] env[61947]: DEBUG nova.network.neutron [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Updating instance_info_cache with network_info: [{"id": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "address": "fa:16:3e:48:bf:c8", "network": {"id": "e87108e8-2032-4242-9f47-bdc532e5913c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2020100740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f1469b7f9a7f4d74ba0580b72c967feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdee94f-08", "ovs_interfaceid": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.415902] env[61947]: DEBUG nova.network.neutron [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.770046] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5282a49e-248c-6aae-d5d6-31ca5bf610d4, 'name': SearchDatastore_Task, 'duration_secs': 0.083614} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.770046] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.770046] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1085.770046] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32425cb9-a6a8-4e98-8f34-90c05c77aa68 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.775656] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1085.775656] env[61947]: value = "task-1224984" [ 1085.775656] env[61947]: _type = "Task" [ 1085.775656] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.786772] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.890771] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.899636] env[61947]: DEBUG nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Received event network-vif-plugged-2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.899781] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Acquiring lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.900017] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.900244] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.900800] env[61947]: DEBUG nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] No waiting events found dispatching network-vif-plugged-2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1085.900800] env[61947]: WARNING nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Received unexpected event network-vif-plugged-2fdee94f-08e4-4c2d-8461-35602d8e5942 for instance with vm_state building and task_state spawning. [ 1085.900800] env[61947]: DEBUG nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Received event network-changed-2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.900976] env[61947]: DEBUG nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Refreshing instance network info cache due to event network-changed-2fdee94f-08e4-4c2d-8461-35602d8e5942. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1085.901312] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Acquiring lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.917783] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Releasing lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.918096] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Instance network_info: |[{"id": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "address": "fa:16:3e:48:bf:c8", "network": {"id": "e87108e8-2032-4242-9f47-bdc532e5913c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2020100740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f1469b7f9a7f4d74ba0580b72c967feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdee94f-08", "ovs_interfaceid": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1085.918469] env[61947]: INFO nova.compute.manager [-] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Took 1.66 seconds to deallocate network for instance. [ 1085.918777] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Acquired lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.918964] env[61947]: DEBUG nova.network.neutron [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Refreshing network info cache for port 2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1085.920292] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:bf:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fdee94f-08e4-4c2d-8461-35602d8e5942', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.928642] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Creating folder: Project (f1469b7f9a7f4d74ba0580b72c967feb). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1085.931595] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db98989f-01bd-465f-b067-3dfdd27345cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.948191] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Created folder: Project (f1469b7f9a7f4d74ba0580b72c967feb) in parent group-v264556. [ 1085.948444] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Creating folder: Instances. Parent ref: group-v264740. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1085.948699] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-637ae532-6bca-4c07-adec-1f135949db33 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.960769] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Created folder: Instances in parent group-v264740. [ 1085.960769] env[61947]: DEBUG oslo.service.loopingcall [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.960975] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1085.961039] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2ce7edd-9c74-4fbe-b9e0-b1846a8dd03b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.987850] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.987850] env[61947]: value = "task-1224987" [ 1085.987850] env[61947]: _type = "Task" [ 1085.987850] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.998830] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224987, 'name': CreateVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.108522] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.287659] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.391705] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.438215] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.438506] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.438734] env[61947]: DEBUG nova.objects.instance [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'resources' on Instance uuid 1faa141b-ddf6-4fc3-bb83-e77efa022f81 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.497953] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1224987, 'name': CreateVM_Task, 'duration_secs': 0.34845} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.500610] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1086.501452] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.501628] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.501959] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.502712] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7ace3b9-2857-4ce2-8ca9-85e30b3bdffd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.507403] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1086.507403] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520f8d10-4000-cd32-6987-6d74f3ea9860" [ 1086.507403] env[61947]: _type = "Task" [ 1086.507403] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.516844] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520f8d10-4000-cd32-6987-6d74f3ea9860, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.610969] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.611271] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1086.611486] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.611646] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.611793] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.611941] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.612102] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.612263] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.612412] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1086.612564] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.643922] env[61947]: DEBUG nova.network.neutron [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Updated VIF entry in instance network info cache for port 2fdee94f-08e4-4c2d-8461-35602d8e5942. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1086.644470] env[61947]: DEBUG nova.network.neutron [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Updating instance_info_cache with network_info: [{"id": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "address": "fa:16:3e:48:bf:c8", "network": {"id": "e87108e8-2032-4242-9f47-bdc532e5913c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2020100740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f1469b7f9a7f4d74ba0580b72c967feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdee94f-08", "ovs_interfaceid": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.787115] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.891557] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.022955] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520f8d10-4000-cd32-6987-6d74f3ea9860, 'name': SearchDatastore_Task, 'duration_secs': 0.083832} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.022955] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.022955] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.022955] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.022955] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.022955] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.022955] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffe98685-6b95-41e7-8b48-1da8960060d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.038346] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.038554] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.039402] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e680a2e2-62a3-4882-b3af-41dd8b14a630 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.044891] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1087.044891] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ac82e-5eb3-7ab0-8f1b-4cb72753df1d" [ 1087.044891] env[61947]: _type = "Task" [ 1087.044891] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.049622] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c8b29a-0551-4654-9e50-4f1b02080341 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.057419] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ac82e-5eb3-7ab0-8f1b-4cb72753df1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.060556] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac08713-a69e-43be-8f59-03499bd955bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.093835] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2a55e8-3f78-4b3e-bd7d-919515930b28 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.101615] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b10f219-65fe-428c-978a-c96e2d5fbaeb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.116387] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.116981] env[61947]: DEBUG nova.compute.provider_tree [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.147384] env[61947]: DEBUG oslo_concurrency.lockutils [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] Releasing lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.147752] env[61947]: DEBUG nova.compute.manager [req-ca73726e-5762-4efd-b5a2-f7a3f24e9a05 req-055297ae-37d0-493f-84ac-f51887f65189 service nova] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Received event network-vif-deleted-7af3de27-3309-4b9e-bf96-5615d232914f {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.287661] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224984, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.393174] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224983, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.880543} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.393531] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a049b242-eba1-46bb-8b53-de799367d098/a049b242-eba1-46bb-8b53-de799367d098.vmdk to [datastore2] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1087.394474] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32668712-d83c-4e80-a2a4-72b2de6f86d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.419140] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.419566] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33566e37-189c-4c9e-a5ef-e5303045fd1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.441340] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1087.441340] env[61947]: value = "task-1224988" [ 1087.441340] env[61947]: _type = "Task" [ 1087.441340] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.450788] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224988, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.556115] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528ac82e-5eb3-7ab0-8f1b-4cb72753df1d, 'name': SearchDatastore_Task, 'duration_secs': 0.082618} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.557018] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf4a4809-d105-4435-a264-791f9726b294 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.562442] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1087.562442] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52193b61-5400-d577-f0bb-9ae5af571f72" [ 1087.562442] env[61947]: _type = "Task" [ 1087.562442] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.570205] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52193b61-5400-d577-f0bb-9ae5af571f72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.620422] env[61947]: DEBUG nova.scheduler.client.report [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.787885] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224984, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.783865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.788349] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1087.788601] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.788905] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c85546a0-af6c-4998-9b55-8d97cde2148e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.795526] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1087.795526] env[61947]: value = "task-1224989" [ 1087.795526] env[61947]: _type = "Task" [ 1087.795526] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.802798] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.953639] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224988, 'name': ReconfigVM_Task, 'duration_secs': 0.343046} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.954073] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 8351ae98-2d33-4f76-a3c1-937738905761/8351ae98-2d33-4f76-a3c1-937738905761.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.954929] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7819a11a-19db-4930-8239-53b73ebd2fa2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.961317] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1087.961317] env[61947]: value = "task-1224990" [ 1087.961317] env[61947]: _type = "Task" [ 1087.961317] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.973314] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224990, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.073558] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52193b61-5400-d577-f0bb-9ae5af571f72, 'name': SearchDatastore_Task, 'duration_secs': 0.00781} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.073860] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.074155] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] b12123c2-5c7b-45fd-81d5-c09abcf8afe4/b12123c2-5c7b-45fd-81d5-c09abcf8afe4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.074432] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da78ac30-9a7c-4639-9e45-bee383ff521b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.080990] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1088.080990] env[61947]: value = "task-1224991" [ 1088.080990] env[61947]: _type = "Task" [ 1088.080990] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.088512] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.125711] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.128214] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.012s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.128489] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.128618] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1088.129578] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0342d38-6531-4bb0-b6df-e00e6a866164 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.139073] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e782a3a-3dff-4118-bf18-2f6026bd6e89 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.154382] env[61947]: INFO nova.scheduler.client.report [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted allocations for instance 1faa141b-ddf6-4fc3-bb83-e77efa022f81 [ 1088.155848] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccd3f94-999f-4165-a0af-4ed6dd5f6b79 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.164991] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16e69b3-ff81-42e7-9c40-a8ffcdf4727e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.199733] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180170MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1088.199906] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.200136] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.305534] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.292175} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.305829] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.306663] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e598647-6514-4337-9865-72906eeeb74e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.333231] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.333592] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9e3b798-dee2-42fe-8428-2cf046a1860e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.355256] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1088.355256] env[61947]: value = "task-1224992" [ 1088.355256] env[61947]: _type = "Task" [ 1088.355256] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.365056] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.471661] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224990, 'name': Rename_Task, 'duration_secs': 0.140083} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.471963] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1088.472270] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4097fc00-dd85-49cd-91e7-184c3fb7d358 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.479823] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1088.479823] env[61947]: value = "task-1224993" [ 1088.479823] env[61947]: _type = "Task" [ 1088.479823] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.488491] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.591370] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427049} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.591691] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] b12123c2-5c7b-45fd-81d5-c09abcf8afe4/b12123c2-5c7b-45fd-81d5-c09abcf8afe4.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1088.591928] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1088.592222] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c2f01b1-57ee-43fc-984b-bc02fd3d4f15 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.598290] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1088.598290] env[61947]: value = "task-1224994" [ 1088.598290] env[61947]: _type = "Task" [ 1088.598290] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.606021] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.664714] env[61947]: DEBUG oslo_concurrency.lockutils [None req-7c50324b-2af8-4cf4-bef5-88ca9354a6d2 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "1faa141b-ddf6-4fc3-bb83-e77efa022f81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.538s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.865315] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224992, 'name': ReconfigVM_Task, 'duration_secs': 0.29544} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.865612] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.866234] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de2e47b2-ff43-46f1-832c-0674c4472782 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.871955] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1088.871955] env[61947]: value = "task-1224995" [ 1088.871955] env[61947]: _type = "Task" [ 1088.871955] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.879890] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224995, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.990020] env[61947]: DEBUG oslo_vmware.api [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1224993, 'name': PowerOnVM_Task, 'duration_secs': 0.440728} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.990307] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1089.104350] env[61947]: DEBUG nova.compute.manager [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.105184] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a258b7-cd2f-4fcf-aea2-c104199110ba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.113361] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.215904} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.115093] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.118795] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89738650-dda7-4ed1-9136-b9435c006f8b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.140745] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] b12123c2-5c7b-45fd-81d5-c09abcf8afe4/b12123c2-5c7b-45fd-81d5-c09abcf8afe4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.141443] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b27a553-6ef4-42d1-8593-27829ec03e55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.160935] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1089.160935] env[61947]: value = "task-1224996" [ 1089.160935] env[61947]: _type = "Task" [ 1089.160935] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.169414] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.225175] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f2bcec82-2730-45bf-b54f-951ba389ae36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.225484] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 8351ae98-2d33-4f76-a3c1-937738905761 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.225484] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.225579] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b12123c2-5c7b-45fd-81d5-c09abcf8afe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1089.225790] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1089.225883] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1089.291390] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010197d8-680e-41d0-9e17-e1c047e27cdf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.300226] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedab09f-dccf-44cc-ab94-10373a23793a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.331065] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e301575-db48-4e0a-b202-cb1a07783477 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.339286] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d418ae30-782c-40a8-b463-75a48f796b3d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.353283] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.382226] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224995, 'name': Rename_Task, 'duration_secs': 0.153655} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.382516] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1089.382813] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b2ba40e-6744-4264-8c3c-d10d4dca1f2f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.388211] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1089.388211] env[61947]: value = "task-1224997" [ 1089.388211] env[61947]: _type = "Task" [ 1089.388211] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.395453] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.631905] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1482307f-ade9-4859-971b-ace375bab04e tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.834s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.675311] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224996, 'name': ReconfigVM_Task, 'duration_secs': 0.307568} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.675311] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Reconfigured VM instance instance-00000065 to attach disk [datastore2] b12123c2-5c7b-45fd-81d5-c09abcf8afe4/b12123c2-5c7b-45fd-81d5-c09abcf8afe4.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.675987] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f710173-ebc8-4929-9d05-b9faa30f054d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.683836] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1089.683836] env[61947]: value = "task-1224998" [ 1089.683836] env[61947]: _type = "Task" [ 1089.683836] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.693910] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224998, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.858575] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.899740] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224997, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.194970] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224998, 'name': Rename_Task, 'duration_secs': 0.136106} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.195357] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.195667] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6c13113-68ed-43a0-b339-2cee97b7aceb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.203834] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1090.203834] env[61947]: value = "task-1224999" [ 1090.203834] env[61947]: _type = "Task" [ 1090.203834] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.213372] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224999, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.362953] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1090.363198] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.163s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.398373] env[61947]: DEBUG oslo_vmware.api [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1224997, 'name': PowerOnVM_Task, 'duration_secs': 0.844715} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.398664] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.398872] env[61947]: INFO nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1090.399067] env[61947]: DEBUG nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.399961] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e515ae6a-0848-4b7d-aefb-cf77c3bea819 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.714222] env[61947]: DEBUG oslo_vmware.api [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1224999, 'name': PowerOnVM_Task, 'duration_secs': 0.497854} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.714501] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.714711] env[61947]: INFO nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Took 6.95 seconds to spawn the instance on the hypervisor. [ 1090.714890] env[61947]: DEBUG nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.715732] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf413e36-b043-46f5-8861-f7596b603351 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.916908] env[61947]: INFO nova.compute.manager [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Took 13.58 seconds to build instance. [ 1090.926660] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84885d50-b783-40c5-b860-1377b0cde203 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.934008] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Suspending the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1090.934269] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4f996838-cf07-4775-9327-0e5e2b5b4fb4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.940139] env[61947]: DEBUG oslo_vmware.api [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1090.940139] env[61947]: value = "task-1225000" [ 1090.940139] env[61947]: _type = "Task" [ 1090.940139] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.947961] env[61947]: DEBUG oslo_vmware.api [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225000, 'name': SuspendVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.234246] env[61947]: INFO nova.compute.manager [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Took 13.55 seconds to build instance. [ 1091.419072] env[61947]: DEBUG oslo_concurrency.lockutils [None req-93142ee1-db9b-4045-803f-61fb6b805744 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.095s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.419346] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.419560] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.451444] env[61947]: DEBUG oslo_vmware.api [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225000, 'name': SuspendVM_Task} progress is 66%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.736611] env[61947]: DEBUG oslo_concurrency.lockutils [None req-10651387-c82b-4c29-9ae4-98d2cebf3a7b tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.063s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.923852] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1091.951433] env[61947]: DEBUG oslo_vmware.api [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225000, 'name': SuspendVM_Task, 'duration_secs': 0.655121} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.951857] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Suspended the VM {{(pid=61947) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1091.952174] env[61947]: DEBUG nova.compute.manager [None req-66bf9730-1b2f-4a8d-9213-267d57be087d tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.953055] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7beaf73d-0600-400f-be45-d36ba5a4d60a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.038511] env[61947]: DEBUG nova.compute.manager [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.038738] env[61947]: DEBUG nova.compute.manager [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing instance network info cache due to event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1092.038966] env[61947]: DEBUG oslo_concurrency.lockutils [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.039131] env[61947]: DEBUG oslo_concurrency.lockutils [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.039298] env[61947]: DEBUG nova.network.neutron [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.152849] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.153145] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.153439] env[61947]: INFO nova.compute.manager [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Rebooting instance [ 1092.444155] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.444496] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.446106] env[61947]: INFO nova.compute.claims [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.679076] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.679403] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquired lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.679621] env[61947]: DEBUG nova.network.neutron [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.761700] env[61947]: DEBUG nova.network.neutron [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updated VIF entry in instance network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1092.762108] env[61947]: DEBUG nova.network.neutron [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.264854] env[61947]: DEBUG oslo_concurrency.lockutils [req-f2fe4e9d-6457-449c-92c6-f953fff862d0 req-ff366a69-16b6-4a0f-a4d2-4dae6458b01e service nova] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.531109] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfc5859-c98f-4b0a-ad40-0f2b59cac724 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.538584] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6b617a-e896-41d5-82bf-e9e867812d81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.570462] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdefb0bd-af6a-4f41-a0fd-1b40ffd02991 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.577870] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e55c1d-377a-4b41-baa5-289ce3dbf72c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.590691] env[61947]: DEBUG nova.compute.provider_tree [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.595694] env[61947]: INFO nova.compute.manager [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Resuming [ 1093.597606] env[61947]: DEBUG nova.objects.instance [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'flavor' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.604778] env[61947]: DEBUG nova.network.neutron [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Updating instance_info_cache with network_info: [{"id": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "address": "fa:16:3e:48:bf:c8", "network": {"id": "e87108e8-2032-4242-9f47-bdc532e5913c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2020100740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f1469b7f9a7f4d74ba0580b72c967feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdee94f-08", "ovs_interfaceid": "2fdee94f-08e4-4c2d-8461-35602d8e5942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.093785] env[61947]: DEBUG nova.scheduler.client.report [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.107230] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Releasing lock "refresh_cache-b12123c2-5c7b-45fd-81d5-c09abcf8afe4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.108989] env[61947]: DEBUG nova.compute.manager [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1094.109901] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ab2038-a33c-4c11-b288-89374456c7a6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.599141] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.154s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.599736] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1094.604607] env[61947]: DEBUG oslo_concurrency.lockutils [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.604776] env[61947]: DEBUG oslo_concurrency.lockutils [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquired lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.604947] env[61947]: DEBUG nova.network.neutron [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1095.105443] env[61947]: DEBUG nova.compute.utils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1095.106788] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1095.106967] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1095.125524] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed56254-126a-4a8b-87ff-2b1e164c36bb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.133332] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Doing hard reboot of VM {{(pid=61947) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1095.133628] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-bd7be158-e335-4312-80a6-ae3e02c2c901 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.140486] env[61947]: DEBUG oslo_vmware.api [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1095.140486] env[61947]: value = "task-1225001" [ 1095.140486] env[61947]: _type = "Task" [ 1095.140486] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.149317] env[61947]: DEBUG oslo_vmware.api [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225001, 'name': ResetVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.196180] env[61947]: DEBUG nova.policy [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '897eb27f633a4a65bd500448ede60c1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7901e6bbfd1e4832a556d1a2b58d9850', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1095.492314] env[61947]: DEBUG nova.network.neutron [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [{"id": "487b8623-6225-453d-9be0-1dd6e1241b97", "address": "fa:16:3e:2a:65:05", "network": {"id": "e432e9c4-09dc-436c-94ef-8992575a6370", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-94781543-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d76ebc8875f4b03bc768320b44b11d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487b8623-62", "ovs_interfaceid": "487b8623-6225-453d-9be0-1dd6e1241b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.611903] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1095.616401] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Successfully created port: 746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1095.651615] env[61947]: DEBUG oslo_vmware.api [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225001, 'name': ResetVM_Task, 'duration_secs': 0.102979} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.652900] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Did hard reboot of VM {{(pid=61947) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1095.653233] env[61947]: DEBUG nova.compute.manager [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.654432] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9caf4e-075c-42e2-9d99-2ae89765f918 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.995684] env[61947]: DEBUG oslo_concurrency.lockutils [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Releasing lock "refresh_cache-8351ae98-2d33-4f76-a3c1-937738905761" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.996661] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa54403-c166-4bc0-8c1d-44de641748ce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.004254] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Resuming the VM {{(pid=61947) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1096.004510] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2f00ed4-4c11-4e30-a395-7973fe5a2de1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.010521] env[61947]: DEBUG oslo_vmware.api [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1096.010521] env[61947]: value = "task-1225002" [ 1096.010521] env[61947]: _type = "Task" [ 1096.010521] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.017667] env[61947]: DEBUG oslo_vmware.api [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225002, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.170087] env[61947]: DEBUG oslo_concurrency.lockutils [None req-40d8a912-f07f-4b7e-838a-a913ad8540a7 tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.017s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.525024] env[61947]: DEBUG oslo_vmware.api [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225002, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.621341] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1096.654223] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1096.654536] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1096.654736] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1096.654931] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1096.655096] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1096.655252] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1096.655531] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1096.655721] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1096.655926] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1096.656120] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1096.656303] env[61947]: DEBUG nova.virt.hardware [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.657278] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7c271d-0f4c-4cd3-83f0-e09ef5a3e915 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.667047] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800e96d7-1099-4723-b9b8-dbfd098b9f50 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.001126] env[61947]: DEBUG nova.compute.manager [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Received event network-vif-plugged-746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1097.001464] env[61947]: DEBUG oslo_concurrency.lockutils [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.001763] env[61947]: DEBUG oslo_concurrency.lockutils [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.002017] env[61947]: DEBUG oslo_concurrency.lockutils [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.002336] env[61947]: DEBUG nova.compute.manager [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] No waiting events found dispatching network-vif-plugged-746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1097.002571] env[61947]: WARNING nova.compute.manager [req-0dadec6c-c81b-4eb2-bb45-a0d4bdd2bb13 req-f243b4fd-1a53-41a3-adb2-95fee95985ee service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Received unexpected event network-vif-plugged-746ecfed-7986-4f58-87a1-1bdddd80d0b2 for instance with vm_state building and task_state spawning. [ 1097.022029] env[61947]: DEBUG oslo_vmware.api [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225002, 'name': PowerOnVM_Task, 'duration_secs': 0.520614} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.022317] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Resumed the VM {{(pid=61947) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1097.022573] env[61947]: DEBUG nova.compute.manager [None req-53e8ec6e-d4c1-48ac-bb8b-d7e9c9579a99 tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.023411] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3369e0f8-092a-4f93-9d83-fb742d64e549 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.099065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.099065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.099065] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.099315] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.099428] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.101628] env[61947]: INFO nova.compute.manager [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Terminating instance [ 1097.103410] env[61947]: DEBUG nova.compute.manager [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.103619] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1097.104467] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1377dac7-3720-4719-bc98-de72469a980f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.112557] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1097.112793] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48aa777c-19e9-4654-a956-d8893afda413 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.119235] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1097.119235] env[61947]: value = "task-1225003" [ 1097.119235] env[61947]: _type = "Task" [ 1097.119235] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.126658] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.537770] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Successfully updated port: 746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1097.560489] env[61947]: DEBUG nova.compute.manager [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Received event network-changed-746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1097.560489] env[61947]: DEBUG nova.compute.manager [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Refreshing instance network info cache due to event network-changed-746ecfed-7986-4f58-87a1-1bdddd80d0b2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1097.560489] env[61947]: DEBUG oslo_concurrency.lockutils [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] Acquiring lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.560888] env[61947]: DEBUG oslo_concurrency.lockutils [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] Acquired lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.560888] env[61947]: DEBUG nova.network.neutron [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Refreshing network info cache for port 746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1097.630715] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225003, 'name': PowerOffVM_Task, 'duration_secs': 0.196923} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.631064] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1097.631192] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1097.631466] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4cfb9c99-7796-4f47-b3b1-bf0b959d894a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.692146] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1097.692403] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1097.692646] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Deleting the datastore file [datastore2] b12123c2-5c7b-45fd-81d5-c09abcf8afe4 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.692955] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee4dc0ad-d748-48ab-84d4-cd27bbc76f63 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.699751] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for the task: (returnval){ [ 1097.699751] env[61947]: value = "task-1225005" [ 1097.699751] env[61947]: _type = "Task" [ 1097.699751] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.708120] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225005, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.040475] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.092280] env[61947]: DEBUG nova.network.neutron [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1098.211449] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225005, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.219023] env[61947]: DEBUG nova.network.neutron [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.710591] env[61947]: DEBUG oslo_vmware.api [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Task: {'id': task-1225005, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.803322} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.710860] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.711057] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1098.711243] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1098.711419] env[61947]: INFO nova.compute.manager [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1098.711691] env[61947]: DEBUG oslo.service.loopingcall [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.711893] env[61947]: DEBUG nova.compute.manager [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.711999] env[61947]: DEBUG nova.network.neutron [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.721160] env[61947]: DEBUG oslo_concurrency.lockutils [req-c77af2ab-af96-44a4-b0e3-4413e3807cba req-ec5981d5-3959-497d-b56b-b6e25c918713 service nova] Releasing lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.721516] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.721685] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1099.250989] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1099.386171] env[61947]: DEBUG nova.network.neutron [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating instance_info_cache with network_info: [{"id": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "address": "fa:16:3e:b6:59:ca", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746ecfed-79", "ovs_interfaceid": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.500186] env[61947]: DEBUG nova.network.neutron [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.586171] env[61947]: DEBUG nova.compute.manager [req-5bf8c10d-3c1d-475b-82f1-40571d4addaf req-5fe21f6a-a796-4491-a987-b9ea62e384d0 service nova] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Received event network-vif-deleted-2fdee94f-08e4-4c2d-8461-35602d8e5942 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.889053] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.889456] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Instance network_info: |[{"id": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "address": "fa:16:3e:b6:59:ca", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746ecfed-79", "ovs_interfaceid": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1099.889855] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:59:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '746ecfed-7986-4f58-87a1-1bdddd80d0b2', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1099.897321] env[61947]: DEBUG oslo.service.loopingcall [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.897532] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1099.897757] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10987710-098c-40d9-bb32-455d35158e5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.917822] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1099.917822] env[61947]: value = "task-1225006" [ 1099.917822] env[61947]: _type = "Task" [ 1099.917822] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.925039] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225006, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.002813] env[61947]: INFO nova.compute.manager [-] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Took 1.29 seconds to deallocate network for instance. [ 1100.428267] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225006, 'name': CreateVM_Task, 'duration_secs': 0.292369} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.428453] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1100.429374] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.429581] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.429925] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1100.430197] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55575026-16de-464c-b7ae-6e9945a44875 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.434716] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1100.434716] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52664ca8-4f83-26ed-3801-44d343ae32f1" [ 1100.434716] env[61947]: _type = "Task" [ 1100.434716] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.442775] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52664ca8-4f83-26ed-3801-44d343ae32f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.509159] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.509529] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.509830] env[61947]: DEBUG nova.objects.instance [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lazy-loading 'resources' on Instance uuid b12123c2-5c7b-45fd-81d5-c09abcf8afe4 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.945409] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52664ca8-4f83-26ed-3801-44d343ae32f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010184} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.945762] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.946086] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1100.946368] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.946561] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.946785] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1100.947088] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-899be7dd-e49b-4e1f-bd1a-487468390d77 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.954973] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1100.955194] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1100.955977] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4148f92-0069-4e93-9a3f-62d62efd0495 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.960771] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1100.960771] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281d37a-3a15-f82b-50a3-a71310deb039" [ 1100.960771] env[61947]: _type = "Task" [ 1100.960771] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.968077] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281d37a-3a15-f82b-50a3-a71310deb039, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.084218] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4be7ef-9f31-4f97-8b89-b68d8c4703c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.091792] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b48581b-4d01-4c6f-beb0-4d77adc27e2a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.121718] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb21640-e0bd-4063-a269-b8a8f5e37215 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.128436] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47e4626-0d82-4b49-9361-5f584cbdedd1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.140971] env[61947]: DEBUG nova.compute.provider_tree [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.471373] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5281d37a-3a15-f82b-50a3-a71310deb039, 'name': SearchDatastore_Task, 'duration_secs': 0.008105} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.472169] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d2f9553-bb49-4692-a331-8e182029c9a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.477171] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1101.477171] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7ca6c-c9c5-6825-b55c-a37be3f44411" [ 1101.477171] env[61947]: _type = "Task" [ 1101.477171] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.484523] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7ca6c-c9c5-6825-b55c-a37be3f44411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.644456] env[61947]: DEBUG nova.scheduler.client.report [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.987898] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52e7ca6c-c9c5-6825-b55c-a37be3f44411, 'name': SearchDatastore_Task, 'duration_secs': 0.009644} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.988280] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.988481] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b9a31c8d-da69-45a6-9921-3a587d212f3d/b9a31c8d-da69-45a6-9921-3a587d212f3d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1101.988741] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d09c7b54-bce5-4388-832a-fb9cdadd5504 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.995298] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1101.995298] env[61947]: value = "task-1225007" [ 1101.995298] env[61947]: _type = "Task" [ 1101.995298] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.002778] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.150159] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.163630] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.164059] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.164286] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "8351ae98-2d33-4f76-a3c1-937738905761-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.164498] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.164805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.167166] env[61947]: INFO nova.compute.manager [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Terminating instance [ 1102.170677] env[61947]: DEBUG nova.compute.manager [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.170912] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1102.171815] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52951dc4-e063-41cf-89c7-27e99458836e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.176157] env[61947]: INFO nova.scheduler.client.report [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Deleted allocations for instance b12123c2-5c7b-45fd-81d5-c09abcf8afe4 [ 1102.185022] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.185385] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2608fe94-e1ab-4488-9e53-f79f40452aea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.193144] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1102.193144] env[61947]: value = "task-1225008" [ 1102.193144] env[61947]: _type = "Task" [ 1102.193144] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.204913] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.504774] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432304} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.505847] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] b9a31c8d-da69-45a6-9921-3a587d212f3d/b9a31c8d-da69-45a6-9921-3a587d212f3d.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1102.505847] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1102.505847] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-997a62e1-c17f-4a38-a95e-b8f5814e4015 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.511326] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1102.511326] env[61947]: value = "task-1225009" [ 1102.511326] env[61947]: _type = "Task" [ 1102.511326] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.518093] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.685453] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2624ca8f-85dd-4273-a0a4-604eebd92ede tempest-InstanceActionsTestJSON-1841975819 tempest-InstanceActionsTestJSON-1841975819-project-member] Lock "b12123c2-5c7b-45fd-81d5-c09abcf8afe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.587s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.703019] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225008, 'name': PowerOffVM_Task, 'duration_secs': 0.266693} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.703291] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.703464] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1102.703701] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2fa53d49-c7c4-40fe-8bbb-1a3686a62d60 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.826056] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.826056] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.826056] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleting the datastore file [datastore2] 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.826404] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3873f3d4-e809-4cdf-9faf-6337229f5014 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.831976] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for the task: (returnval){ [ 1102.831976] env[61947]: value = "task-1225011" [ 1102.831976] env[61947]: _type = "Task" [ 1102.831976] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.838958] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.020770] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062217} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.021100] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1103.021833] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c163e6-9afa-4fc6-8cac-478e3b2b8438 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.043265] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] b9a31c8d-da69-45a6-9921-3a587d212f3d/b9a31c8d-da69-45a6-9921-3a587d212f3d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.043514] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5670503-b2c3-4a94-a26e-64100e832af8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.061578] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1103.061578] env[61947]: value = "task-1225012" [ 1103.061578] env[61947]: _type = "Task" [ 1103.061578] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.070507] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225012, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.342391] env[61947]: DEBUG oslo_vmware.api [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Task: {'id': task-1225011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139589} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.342698] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.342920] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1103.343121] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1103.343622] env[61947]: INFO nova.compute.manager [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1103.343896] env[61947]: DEBUG oslo.service.loopingcall [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.344101] env[61947]: DEBUG nova.compute.manager [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.344200] env[61947]: DEBUG nova.network.neutron [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1103.567489] env[61947]: DEBUG nova.compute.manager [req-0ac6d3a5-f56b-47a4-a313-6874454f0174 req-bc58b93e-e09f-427c-8e7e-84757cbb3feb service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Received event network-vif-deleted-487b8623-6225-453d-9be0-1dd6e1241b97 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1103.567699] env[61947]: INFO nova.compute.manager [req-0ac6d3a5-f56b-47a4-a313-6874454f0174 req-bc58b93e-e09f-427c-8e7e-84757cbb3feb service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Neutron deleted interface 487b8623-6225-453d-9be0-1dd6e1241b97; detaching it from the instance and deleting it from the info cache [ 1103.567905] env[61947]: DEBUG nova.network.neutron [req-0ac6d3a5-f56b-47a4-a313-6874454f0174 req-bc58b93e-e09f-427c-8e7e-84757cbb3feb service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.574318] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225012, 'name': ReconfigVM_Task, 'duration_secs': 0.265808} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.574635] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfigured VM instance instance-00000066 to attach disk [datastore1] b9a31c8d-da69-45a6-9921-3a587d212f3d/b9a31c8d-da69-45a6-9921-3a587d212f3d.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1103.575404] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee2a43c6-6d93-48cb-ad11-45e0857b7cf0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.581233] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1103.581233] env[61947]: value = "task-1225013" [ 1103.581233] env[61947]: _type = "Task" [ 1103.581233] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.590556] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225013, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.044446] env[61947]: DEBUG nova.network.neutron [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.070548] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c204761a-87df-4e5a-b5de-42e535eb3d14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.079657] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0ec3fd-cf0a-4621-8452-b4d4badc7fa7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.100381] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225013, 'name': Rename_Task, 'duration_secs': 0.458894} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.100463] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1104.100811] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8a35291-d7a2-428d-b6b8-f7a316029742 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.108672] env[61947]: DEBUG nova.compute.manager [req-0ac6d3a5-f56b-47a4-a313-6874454f0174 req-bc58b93e-e09f-427c-8e7e-84757cbb3feb service nova] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Detach interface failed, port_id=487b8623-6225-453d-9be0-1dd6e1241b97, reason: Instance 8351ae98-2d33-4f76-a3c1-937738905761 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1104.113462] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1104.113462] env[61947]: value = "task-1225014" [ 1104.113462] env[61947]: _type = "Task" [ 1104.113462] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.120617] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225014, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.547580] env[61947]: INFO nova.compute.manager [-] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Took 1.20 seconds to deallocate network for instance. [ 1104.623542] env[61947]: DEBUG oslo_vmware.api [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225014, 'name': PowerOnVM_Task, 'duration_secs': 0.483762} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.623817] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1104.624155] env[61947]: INFO nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1104.624357] env[61947]: DEBUG nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.625568] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b848e405-8631-48a1-bfbf-cae9a7c6448c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.056215] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.056510] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.056843] env[61947]: DEBUG nova.objects.instance [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lazy-loading 'resources' on Instance uuid 8351ae98-2d33-4f76-a3c1-937738905761 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.143075] env[61947]: INFO nova.compute.manager [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Took 12.72 seconds to build instance. [ 1105.622802] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02a8cf2-c7b1-4108-9a38-ddb22d65030c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.630658] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b29dfdf-e717-4fb4-9063-0987b2183bbc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.660905] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ba1273e7-e315-486c-b1b4-805e9797d652 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.241s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.662419] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897ff822-8160-4573-98fc-306dac57597f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.669515] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e0368f-dc6b-4b6f-bb2f-53179d9c3c0c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.682571] env[61947]: DEBUG nova.compute.provider_tree [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.718589] env[61947]: DEBUG nova.compute.manager [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Received event network-changed-746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.718735] env[61947]: DEBUG nova.compute.manager [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Refreshing instance network info cache due to event network-changed-746ecfed-7986-4f58-87a1-1bdddd80d0b2. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1105.718953] env[61947]: DEBUG oslo_concurrency.lockutils [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] Acquiring lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.719115] env[61947]: DEBUG oslo_concurrency.lockutils [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] Acquired lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.719286] env[61947]: DEBUG nova.network.neutron [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Refreshing network info cache for port 746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.185181] env[61947]: DEBUG nova.scheduler.client.report [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.420366] env[61947]: DEBUG nova.network.neutron [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updated VIF entry in instance network info cache for port 746ecfed-7986-4f58-87a1-1bdddd80d0b2. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1106.420732] env[61947]: DEBUG nova.network.neutron [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating instance_info_cache with network_info: [{"id": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "address": "fa:16:3e:b6:59:ca", "network": {"id": "40719252-22b8-4b3d-aa6f-479429b41815", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1991722802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7901e6bbfd1e4832a556d1a2b58d9850", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746ecfed-79", "ovs_interfaceid": "746ecfed-7986-4f58-87a1-1bdddd80d0b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.690433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.709170] env[61947]: INFO nova.scheduler.client.report [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Deleted allocations for instance 8351ae98-2d33-4f76-a3c1-937738905761 [ 1106.923726] env[61947]: DEBUG oslo_concurrency.lockutils [req-7fd740fa-b77d-4804-82ec-79552dd365c2 req-11361a62-10d7-415e-bb42-793f01a36456 service nova] Releasing lock "refresh_cache-b9a31c8d-da69-45a6-9921-3a587d212f3d" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.216913] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3c9ed5d1-e7b7-4ab6-881f-c787b481b1df tempest-ServersNegativeTestJSON-1748569739 tempest-ServersNegativeTestJSON-1748569739-project-member] Lock "8351ae98-2d33-4f76-a3c1-937738905761" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.053s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.351927] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.351927] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.854484] env[61947]: DEBUG nova.compute.utils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.357880] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.415834] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.416249] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.416374] env[61947]: INFO nova.compute.manager [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attaching volume 57b1923c-cd8f-4f51-a82a-4c19800dadc2 to /dev/sdb [ 1113.446600] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66ee50a-e5b3-4cd7-b215-906d25abb3f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.454152] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519f2bbb-5f5b-471a-87cf-bfdf2d6fbd32 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.466643] env[61947]: DEBUG nova.virt.block_device [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating existing volume attachment record: 309fa5a4-3c67-442c-82cc-96abfc05debf {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 1116.385425] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.385707] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.385806] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1116.888990] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Didn't find any instances for network info cache update. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1116.889263] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.889429] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.889586] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.392212] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.392570] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.392686] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.392786] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1117.393746] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34be2d2-a635-4a00-b348-5a7cbfe874fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.401773] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d89bc5-cffb-4f7d-8037-72b269c086c0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.415770] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c754de94-ec1f-4238-be5e-984241af51e7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.421534] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48a92de-fdb7-4fd9-b8f2-5ad7ea5704ea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.450246] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180170MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1117.450389] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.450565] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.009338] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1118.009632] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264744', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'name': 'volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'serial': '57b1923c-cd8f-4f51-a82a-4c19800dadc2'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1118.010645] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2333b0bd-f10f-4997-ac7e-784bd6a17305 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.027063] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd61c3a-0551-4023-850f-e5ea0c06d26e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.050255] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2/volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.050486] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a50ab840-21db-4b33-a892-53564022678f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.067791] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1118.067791] env[61947]: value = "task-1225017" [ 1118.067791] env[61947]: _type = "Task" [ 1118.067791] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.074899] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225017, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.474798] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance f2bcec82-2730-45bf-b54f-951ba389ae36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1118.475126] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1118.475126] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance b9a31c8d-da69-45a6-9921-3a587d212f3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1118.475270] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1118.475409] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1118.518735] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b38e60-fe23-44e0-967c-07774866c301 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.526353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ac92cf-9a0f-4505-b98c-4b4f9a3e56f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.556240] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef24a218-db54-4fd5-8bd9-9e852441c223 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.562719] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d984a0-0ec3-4603-af8b-d035848252d9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.575621] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.583408] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225017, 'name': ReconfigVM_Task, 'duration_secs': 0.315591} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.583668] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2/volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.588717] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-548e44ad-2e6d-4a39-a14f-3fcbd474f166 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.602724] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1118.602724] env[61947]: value = "task-1225018" [ 1118.602724] env[61947]: _type = "Task" [ 1118.602724] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.610166] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225018, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.080837] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.113427] env[61947]: DEBUG oslo_vmware.api [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225018, 'name': ReconfigVM_Task, 'duration_secs': 0.130751} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.113737] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264744', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'name': 'volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'serial': '57b1923c-cd8f-4f51-a82a-4c19800dadc2'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1119.585731] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1119.586100] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.135s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.586213] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.586362] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Cleaning up deleted instances {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1120.101274] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] There are 55 instances to clean {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1120.101644] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: b12123c2-5c7b-45fd-81d5-c09abcf8afe4] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.147463] env[61947]: DEBUG nova.objects.instance [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.604874] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 1faa141b-ddf6-4fc3-bb83-e77efa022f81] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.653435] env[61947]: DEBUG oslo_concurrency.lockutils [None req-3cac74c3-0a4b-4a5b-8a94-7af3ed1fcb99 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.237s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.108163] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: d68e1fc3-0129-432b-adce-30773f5bd6ee] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.467587] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.467839] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.611782] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: a56bb67b-93e4-4ec1-a3c6-b9172a4ed617] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.971289] env[61947]: DEBUG nova.compute.utils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.115209] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 068c9c5e-f7e4-4a5e-896d-a5c348949e07] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1122.474830] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.618681] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 0da87f27-f431-4ebf-ac5e-571605481627] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1123.123071] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4e0a8c79-9445-4474-b516-e697e1854dea] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1123.539374] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.539665] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.539908] env[61947]: INFO nova.compute.manager [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attaching volume 07f144fe-e2ce-4585-ada1-6d40566b3433 to /dev/sdc [ 1123.569803] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac0da87-c26c-4dee-84c3-2a9864c4ebe3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.577516] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ceb8ec0-cc05-46d5-96d8-9aa2b807bcfd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.590082] env[61947]: DEBUG nova.virt.block_device [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating existing volume attachment record: 7c154579-2c0a-404f-9ac6-8042acce9758 {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 1123.625927] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8a4b81d3-d683-4517-806c-88b41ad73c48] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1124.129640] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 1660df2a-b2c5-469f-b602-0014dd3656d5] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1124.633151] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8351ae98-2d33-4f76-a3c1-937738905761] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1125.136871] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 41f62ad6-b11e-4c77-98bc-47e26234112c] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1125.639752] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ebf276db-cc9b-441f-a01b-7f7dc5b83fd3] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1126.143228] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 85210032-a4e2-4f76-96e7-efa2fa6e134e] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1126.647179] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 9378657d-dc54-47e2-b178-1ed3f9c952f2] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1127.150472] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: e137f21e-766d-4b20-9d92-5d1907e2baa3] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1127.653900] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: fcd2e134-4d5f-43d5-a09c-55a49f8ce48b] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1128.132188] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1128.132464] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264745', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'name': 'volume-07f144fe-e2ce-4585-ada1-6d40566b3433', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'serial': '07f144fe-e2ce-4585-ada1-6d40566b3433'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1128.133401] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bce38e-15bd-4eb9-8e7e-6e12872d2cff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.149093] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bc05f5-bf4d-4894-b1fe-d0ba9b12af10 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.168469] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: d2d094c4-0237-48cd-b98b-8d6c49ee5d94] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1128.177411] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-07f144fe-e2ce-4585-ada1-6d40566b3433/volume-07f144fe-e2ce-4585-ada1-6d40566b3433.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.177898] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d89bef70-ad6c-4639-a70a-4c81dad294e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.196479] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1128.196479] env[61947]: value = "task-1225021" [ 1128.196479] env[61947]: _type = "Task" [ 1128.196479] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.205979] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225021, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.678482] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ab21547e-0ddb-4271-967f-bea5056c163a] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1128.705891] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225021, 'name': ReconfigVM_Task, 'duration_secs': 0.332657} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.706198] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-07f144fe-e2ce-4585-ada1-6d40566b3433/volume-07f144fe-e2ce-4585-ada1-6d40566b3433.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.711099] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0dd4af4-652b-475d-a43f-4208f4785eda {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.725521] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1128.725521] env[61947]: value = "task-1225022" [ 1128.725521] env[61947]: _type = "Task" [ 1128.725521] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.733297] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.181811] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 16dd2b82-1925-410b-b3b8-d3e2cf8648b0] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1129.235644] env[61947]: DEBUG oslo_vmware.api [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225022, 'name': ReconfigVM_Task, 'duration_secs': 0.127262} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.235969] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264745', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'name': 'volume-07f144fe-e2ce-4585-ada1-6d40566b3433', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'serial': '07f144fe-e2ce-4585-ada1-6d40566b3433'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1129.527434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.527869] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.528100] env[61947]: INFO nova.compute.manager [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Shelving [ 1129.685170] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: e6435156-fb2e-4912-9587-99eeafeded87] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1130.036027] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1130.036401] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67385c51-3cd1-4b41-b9d1-9613c47daae7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.044230] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1130.044230] env[61947]: value = "task-1225023" [ 1130.044230] env[61947]: _type = "Task" [ 1130.044230] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.052313] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.188400] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: c4647324-7e52-44d4-a4b0-961684cc1c4c] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1130.274836] env[61947]: DEBUG nova.objects.instance [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.556016] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225023, 'name': PowerOffVM_Task, 'duration_secs': 0.19419} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.556383] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1130.557264] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126d41b7-ad36-4ed8-a227-e663074f285e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.583080] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e2c749-b728-4b6a-988a-ab07dc3092e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.692073] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: dcda7cbf-4d1f-4c4e-90bc-9b6aa97145a2] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1130.780073] env[61947]: DEBUG oslo_concurrency.lockutils [None req-9a21c419-15db-497b-8d5d-824cb837e5d3 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.066123] env[61947]: DEBUG oslo_concurrency.lockutils [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.066400] env[61947]: DEBUG oslo_concurrency.lockutils [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.094108] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Creating Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1131.094670] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-680a31e2-316a-4de8-9653-db230964b8b3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.102416] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1131.102416] env[61947]: value = "task-1225024" [ 1131.102416] env[61947]: _type = "Task" [ 1131.102416] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.111091] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225024, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.196836] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 91205e54-6bcb-403c-8308-0f0692fe4fba] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1131.569910] env[61947]: INFO nova.compute.manager [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Detaching volume 57b1923c-cd8f-4f51-a82a-4c19800dadc2 [ 1131.600659] env[61947]: INFO nova.virt.block_device [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attempting to driver detach volume 57b1923c-cd8f-4f51-a82a-4c19800dadc2 from mountpoint /dev/sdb [ 1131.600918] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1131.601123] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264744', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'name': 'volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'serial': '57b1923c-cd8f-4f51-a82a-4c19800dadc2'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1131.601999] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551b8a83-54fe-4505-9b86-3bb062be8e05 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.612770] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225024, 'name': CreateSnapshot_Task, 'duration_secs': 0.389785} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.631521] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Created Snapshot of the VM instance {{(pid=61947) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1131.632311] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40a384d-3f36-49ef-aff6-c29a2daf0ffe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.634985] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52b6f31-0cf7-4426-b792-1a2b5fd2ee65 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.646313] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00da8242-566a-4a76-89a9-ca1ddedb7a8c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.669337] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7358c82-03c2-45b0-9a88-983fbc7cfb21 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.683547] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] The volume has not been displaced from its original location: [datastore2] volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2/volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1131.688713] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1131.689206] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c839ca93-7bb0-416f-93d4-f9092d3d18ff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.701791] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 21bb0270-bc20-4ec1-9599-d676845b0dc7] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1131.709091] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1131.709091] env[61947]: value = "task-1225025" [ 1131.709091] env[61947]: _type = "Task" [ 1131.709091] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.716579] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225025, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.157680] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Creating linked-clone VM from snapshot {{(pid=61947) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1132.157998] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-33ca25f8-1377-4970-972b-5b1927c68a01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.166237] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1132.166237] env[61947]: value = "task-1225026" [ 1132.166237] env[61947]: _type = "Task" [ 1132.166237] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.174525] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225026, 'name': CloneVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.205062] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 055a0b7f-1703-447c-b9d4-0a8efc4f5dbb] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1132.218914] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225025, 'name': ReconfigVM_Task, 'duration_secs': 0.212865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.218914] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1132.223790] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b37e7cf-eb30-4421-a4d4-8bed3e9c304c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.238724] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1132.238724] env[61947]: value = "task-1225027" [ 1132.238724] env[61947]: _type = "Task" [ 1132.238724] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.246432] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.676832] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225026, 'name': CloneVM_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.708534] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: f4d0e4d6-1c77-48e3-9f68-7c0b3a01301b] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1132.750291] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225027, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.176430] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225026, 'name': CloneVM_Task, 'duration_secs': 0.931294} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.176683] env[61947]: INFO nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Created linked-clone VM from snapshot [ 1133.177412] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb330fe-4034-42db-8e64-95f0de5c4f8b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.184309] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Uploading image f02d075c-e79e-4232-8cda-baac34044bf4 {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1133.206633] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1133.206633] env[61947]: value = "vm-264747" [ 1133.206633] env[61947]: _type = "VirtualMachine" [ 1133.206633] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1133.206880] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-06e3dbfb-d63a-4967-aafe-227e0caae99b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.212997] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4f313bb5-5df4-4fb0-9736-95f4a519651d] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1133.214908] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease: (returnval){ [ 1133.214908] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fcdfd9-13e4-76e0-9742-e40c02113cbc" [ 1133.214908] env[61947]: _type = "HttpNfcLease" [ 1133.214908] env[61947]: } obtained for exporting VM: (result){ [ 1133.214908] env[61947]: value = "vm-264747" [ 1133.214908] env[61947]: _type = "VirtualMachine" [ 1133.214908] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1133.215135] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the lease: (returnval){ [ 1133.215135] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fcdfd9-13e4-76e0-9742-e40c02113cbc" [ 1133.215135] env[61947]: _type = "HttpNfcLease" [ 1133.215135] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1133.221275] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1133.221275] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fcdfd9-13e4-76e0-9742-e40c02113cbc" [ 1133.221275] env[61947]: _type = "HttpNfcLease" [ 1133.221275] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1133.248249] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225027, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.716324] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 7f308223-6c3c-4c62-8401-d7a44eeb69d2] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1133.725342] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1133.725342] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fcdfd9-13e4-76e0-9742-e40c02113cbc" [ 1133.725342] env[61947]: _type = "HttpNfcLease" [ 1133.725342] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1133.725613] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1133.725613] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52fcdfd9-13e4-76e0-9742-e40c02113cbc" [ 1133.725613] env[61947]: _type = "HttpNfcLease" [ 1133.725613] env[61947]: }. {{(pid=61947) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1133.726316] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d864e611-21a3-41a5-af0f-8f3781f875cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.732900] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1133.733104] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk for reading. {{(pid=61947) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1133.798212] env[61947]: DEBUG oslo_vmware.api [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225027, 'name': ReconfigVM_Task, 'duration_secs': 1.138603} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.798470] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264744', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'name': 'volume-57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '57b1923c-cd8f-4f51-a82a-4c19800dadc2', 'serial': '57b1923c-cd8f-4f51-a82a-4c19800dadc2'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1133.820057] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-52e3d6fa-cfd4-42d2-80cb-dfe139134aba {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.223172] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: e0eebf59-c84e-4462-b280-d783a04525e2] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1134.341669] env[61947]: DEBUG nova.objects.instance [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.727078] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ee3b9509-5437-41b3-b612-91d148a338b6] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1135.231965] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 4f0b9636-4793-4b1c-8b2b-7f31a15669bb] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1135.342279] env[61947]: DEBUG oslo_concurrency.lockutils [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.351021] env[61947]: DEBUG oslo_concurrency.lockutils [None req-73618bdd-c1e5-4595-9279-46664fd4da18 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.284s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.351403] env[61947]: DEBUG oslo_concurrency.lockutils [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.009s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.736103] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 8c342d64-75f6-4b7d-829d-889eb3b6122f] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1135.854617] env[61947]: INFO nova.compute.manager [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Detaching volume 07f144fe-e2ce-4585-ada1-6d40566b3433 [ 1135.888921] env[61947]: INFO nova.virt.block_device [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Attempting to driver detach volume 07f144fe-e2ce-4585-ada1-6d40566b3433 from mountpoint /dev/sdc [ 1135.889074] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1135.889248] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264745', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'name': 'volume-07f144fe-e2ce-4585-ada1-6d40566b3433', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'serial': '07f144fe-e2ce-4585-ada1-6d40566b3433'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1135.890500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e212864-290f-488d-a36c-5d9dd84c1fc5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.912308] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9191175-0259-444d-8e97-29fd0f23258b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.919098] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6883f911-f177-4b90-b109-f9bd3d52c4f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.939583] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e44dc55-beef-48bf-a806-7e3fd3fb9440 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.954477] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] The volume has not been displaced from its original location: [datastore2] volume-07f144fe-e2ce-4585-ada1-6d40566b3433/volume-07f144fe-e2ce-4585-ada1-6d40566b3433.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1135.959965] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfiguring VM instance instance-00000063 to detach disk 2002 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1135.960377] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f1fa077-3dda-43fe-ad4b-bdedff4324d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.978943] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1135.978943] env[61947]: value = "task-1225029" [ 1135.978943] env[61947]: _type = "Task" [ 1135.978943] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.987232] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225029, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.239908] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 9bc18645-c837-4dab-8e05-fbcf0d40812f] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1136.489614] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225029, 'name': ReconfigVM_Task, 'duration_secs': 0.207989} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.489945] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Reconfigured VM instance instance-00000063 to detach disk 2002 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1136.494719] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaf271d4-168b-4a0c-878b-86918e4b4616 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.510028] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1136.510028] env[61947]: value = "task-1225030" [ 1136.510028] env[61947]: _type = "Task" [ 1136.510028] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.517931] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.743493] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: bd84b03e-ef6d-41e0-b48c-b8e8fc425a36] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1137.021830] env[61947]: DEBUG oslo_vmware.api [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225030, 'name': ReconfigVM_Task, 'duration_secs': 0.127803} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.022101] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264745', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'name': 'volume-07f144fe-e2ce-4585-ada1-6d40566b3433', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f2bcec82-2730-45bf-b54f-951ba389ae36', 'attached_at': '', 'detached_at': '', 'volume_id': '07f144fe-e2ce-4585-ada1-6d40566b3433', 'serial': '07f144fe-e2ce-4585-ada1-6d40566b3433'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1137.246556] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 72818e40-624b-4c04-888b-bb622f7f96d7] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1137.563553] env[61947]: DEBUG nova.objects.instance [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'flavor' on Instance uuid f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.750477] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 7d7f267a-3418-4b9b-9589-c915d5194f0d] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1138.254207] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: eede4f34-5ac4-475c-a74b-a98327f648fd] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1138.571473] env[61947]: DEBUG oslo_concurrency.lockutils [None req-61529e6f-f424-4f34-977f-43d99844e907 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.220s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.757643] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 04e5cc55-71d9-4d3d-95c1-fb1401ab74f8] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1139.260432] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: f64b9f7f-89a9-4fb2-9c76-13cc591b9d51] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1139.745818] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.746068] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.746304] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.746534] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.746677] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.748929] env[61947]: INFO nova.compute.manager [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Terminating instance [ 1139.750799] env[61947]: DEBUG nova.compute.manager [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1139.750954] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1139.751848] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04161676-377b-4e45-b697-0b5d6fb32b7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.760221] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1139.760768] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee75b22b-5f61-4957-9c3d-029478a2458d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.762725] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: f68a8537-5bae-4b50-b0ff-fa9b4a89bd40] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1139.766724] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1139.766724] env[61947]: value = "task-1225031" [ 1139.766724] env[61947]: _type = "Task" [ 1139.766724] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.775781] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.266116] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 74e03575-297e-4e08-9236-98d8be80b546] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1140.277308] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225031, 'name': PowerOffVM_Task, 'duration_secs': 0.215385} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.277592] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1140.277777] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1140.278064] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3145adf-ef78-4441-ab1a-3a3ac8cceef0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.336141] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1140.336389] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1140.336612] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleting the datastore file [datastore2] f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1140.336878] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7513f44e-31b7-4de3-a872-e1b8fa45d520 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.344948] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for the task: (returnval){ [ 1140.344948] env[61947]: value = "task-1225033" [ 1140.344948] env[61947]: _type = "Task" [ 1140.344948] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.352149] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225033, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.771930] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: d74adcc2-75e6-40ac-b3bf-9cd3f654dbdd] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1140.854727] env[61947]: DEBUG oslo_vmware.api [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Task: {'id': task-1225033, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169265} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.855030] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1140.855224] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1140.855414] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1140.855584] env[61947]: INFO nova.compute.manager [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1140.855837] env[61947]: DEBUG oslo.service.loopingcall [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.856053] env[61947]: DEBUG nova.compute.manager [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1140.856147] env[61947]: DEBUG nova.network.neutron [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1141.215987] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1141.216955] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52d2aa5-8f5b-4db8-be6c-917a1d368071 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.223810] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1141.223979] env[61947]: ERROR oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk due to incomplete transfer. [ 1141.224213] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-65e1ba0b-5f21-4d5d-bad3-762689b1f213 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.230496] env[61947]: DEBUG oslo_vmware.rw_handles [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205a00e-982c-0621-e1e1-4caea1ff4b57/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1141.230693] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Uploaded image f02d075c-e79e-4232-8cda-baac34044bf4 to the Glance image server {{(pid=61947) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1141.232974] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Destroying the VM {{(pid=61947) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1141.233224] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3caae78b-502e-4840-8eed-87c2fd95e74e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.238165] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1141.238165] env[61947]: value = "task-1225034" [ 1141.238165] env[61947]: _type = "Task" [ 1141.238165] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.245465] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225034, 'name': Destroy_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.275635] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 7efd65d9-67f9-497d-9e42-51d8973aa695] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1141.301138] env[61947]: DEBUG nova.compute.manager [req-33cae086-4542-41af-a6a5-28149c424b8f req-516a8c18-945b-43da-90cd-1d4338806086 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Received event network-vif-deleted-4e33aa48-55f8-4eea-be2f-d3b1f986b08b {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1141.301138] env[61947]: INFO nova.compute.manager [req-33cae086-4542-41af-a6a5-28149c424b8f req-516a8c18-945b-43da-90cd-1d4338806086 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Neutron deleted interface 4e33aa48-55f8-4eea-be2f-d3b1f986b08b; detaching it from the instance and deleting it from the info cache [ 1141.301389] env[61947]: DEBUG nova.network.neutron [req-33cae086-4542-41af-a6a5-28149c424b8f req-516a8c18-945b-43da-90cd-1d4338806086 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.748192] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225034, 'name': Destroy_Task, 'duration_secs': 0.305837} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.748482] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Destroyed the VM [ 1141.748727] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleting Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1141.748977] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9fc46605-708d-4429-9651-20af460035cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.755247] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1141.755247] env[61947]: value = "task-1225035" [ 1141.755247] env[61947]: _type = "Task" [ 1141.755247] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.762743] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225035, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.778224] env[61947]: DEBUG nova.network.neutron [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.779550] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: fa0eed6e-831f-4e7a-9890-fc5a02609aa4] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1141.803639] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b87b1e0-6ef5-4915-9dd9-64404a5bf38c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.814817] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110b0c8d-fff3-44eb-96d0-ae68f0e3220c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.841971] env[61947]: DEBUG nova.compute.manager [req-33cae086-4542-41af-a6a5-28149c424b8f req-516a8c18-945b-43da-90cd-1d4338806086 service nova] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Detach interface failed, port_id=4e33aa48-55f8-4eea-be2f-d3b1f986b08b, reason: Instance f2bcec82-2730-45bf-b54f-951ba389ae36 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1142.264873] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225035, 'name': RemoveSnapshot_Task, 'duration_secs': 0.330296} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.267046] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleted Snapshot of the VM instance {{(pid=61947) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1142.267046] env[61947]: DEBUG nova.compute.manager [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.267046] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b9bcd4-8d2c-4ffc-adfd-359bf2678466 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.281600] env[61947]: INFO nova.compute.manager [-] [instance: f2bcec82-2730-45bf-b54f-951ba389ae36] Took 1.43 seconds to deallocate network for instance. [ 1142.281950] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: cc1f84d4-0a28-48af-88ac-7d977db1cf34] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1142.777580] env[61947]: INFO nova.compute.manager [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Shelve offloading [ 1142.779276] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.779571] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-167def86-2b5c-468e-9729-985b7e2966ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.785207] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ab027a9d-f51d-4cb6-8aaf-ebd507548657] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1142.789052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.789052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.789052] env[61947]: DEBUG nova.objects.instance [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lazy-loading 'resources' on Instance uuid f2bcec82-2730-45bf-b54f-951ba389ae36 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.789586] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1142.789586] env[61947]: value = "task-1225036" [ 1142.789586] env[61947]: _type = "Task" [ 1142.789586] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.798462] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] VM already powered off {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1142.798670] env[61947]: DEBUG nova.compute.manager [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.799372] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cace2a4-bc65-42f5-a985-80800c8910fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.804752] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.804903] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.805085] env[61947]: DEBUG nova.network.neutron [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1142.891038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.891295] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.293618] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 9f0c816d-e7e4-4ca9-aee7-8a0280bfe0cb] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1143.396042] env[61947]: DEBUG nova.compute.utils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.429362] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44afeda8-c41e-40b7-950c-d151a3ff6911 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.438499] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b64531-63c0-4832-918c-3f0ccb4d9154 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.471206] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828974f0-0460-4ec4-8996-08ce86d92304 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.478243] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3597244c-ad13-4469-9cfb-e5c97b4edb26 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.490762] env[61947]: DEBUG nova.compute.provider_tree [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.535253] env[61947]: DEBUG nova.network.neutron [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.797277] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 7612f830-fed2-4dd7-ba8f-f792e1aa5ac6] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1143.899059] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.993684] env[61947]: DEBUG nova.scheduler.client.report [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.037475] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.248470] env[61947]: DEBUG nova.compute.manager [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-vif-unplugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.248470] env[61947]: DEBUG oslo_concurrency.lockutils [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.248639] env[61947]: DEBUG oslo_concurrency.lockutils [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.248827] env[61947]: DEBUG oslo_concurrency.lockutils [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.248999] env[61947]: DEBUG nova.compute.manager [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] No waiting events found dispatching network-vif-unplugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1144.249184] env[61947]: WARNING nova.compute.manager [req-48e58295-0008-4be0-bf68-c04d2995d7fb req-b55458b5-1821-462c-b1a0-e1e8b70d40c2 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received unexpected event network-vif-unplugged-5c342e49-9cb5-4596-ae93-6dd705d98225 for instance with vm_state shelved and task_state shelving_offloading. [ 1144.301640] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 2fc54703-5438-4c2a-b1e7-77431ce5177d] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1144.329545] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1144.330725] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d483a7-dfba-44ff-873a-31a2e72a59a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.339034] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1144.339259] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81474e63-26e1-4360-8c7e-bd0020de187d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.498342] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.517280] env[61947]: INFO nova.scheduler.client.report [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Deleted allocations for instance f2bcec82-2730-45bf-b54f-951ba389ae36 [ 1144.675094] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1144.675094] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1144.675094] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleting the datastore file [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.675094] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5302628-b76f-4b48-bc20-b80de7c81d4b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.681013] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1144.681013] env[61947]: value = "task-1225038" [ 1144.681013] env[61947]: _type = "Task" [ 1144.681013] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.689152] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.804865] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: b7d84a56-edb6-4d1e-b7e9-294078be79d8] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1144.958129] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.958446] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.958701] env[61947]: INFO nova.compute.manager [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Attaching volume 54893030-1423-4762-aec4-3e6074904507 to /dev/sdb [ 1144.989804] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac01419-b067-4ac9-a5cd-8c4248d0450b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.996806] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8331cb1-52a4-43e3-82bd-cc88337c9802 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.010765] env[61947]: DEBUG nova.virt.block_device [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating existing volume attachment record: c47883a5-6d3c-44d7-baa2-f826c7eda65c {{(pid=61947) _volume_attach /opt/stack/nova/nova/virt/block_device.py:665}} [ 1145.025620] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e0aba794-35d8-4537-b931-8417dd9215f5 tempest-AttachVolumeTestJSON-1814518934 tempest-AttachVolumeTestJSON-1814518934-project-member] Lock "f2bcec82-2730-45bf-b54f-951ba389ae36" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.280s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.190792] env[61947]: DEBUG oslo_vmware.api [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130848} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.191159] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.191417] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1145.191684] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1145.214441] env[61947]: INFO nova.scheduler.client.report [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted allocations for instance 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 [ 1145.308765] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 9a54ca4f-e6ec-4413-b162-fca0cd824e00] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1145.719073] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.719418] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.719662] env[61947]: DEBUG nova.objects.instance [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'resources' on Instance uuid 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.812562] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: c8f37b6f-4f0c-43db-a1ea-3c45d3956f66] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1146.222262] env[61947]: DEBUG nova.objects.instance [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'numa_topology' on Instance uuid 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.277366] env[61947]: DEBUG nova.compute.manager [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1146.277574] env[61947]: DEBUG nova.compute.manager [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing instance network info cache due to event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1146.277791] env[61947]: DEBUG oslo_concurrency.lockutils [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.277936] env[61947]: DEBUG oslo_concurrency.lockutils [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.278489] env[61947]: DEBUG nova.network.neutron [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1146.315132] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: d642bdb3-8e98-4667-9af7-71dae739e6b1] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1146.726047] env[61947]: DEBUG nova.objects.base [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Object Instance<41dc3d63-d455-4f74-9ec8-a7806f25b2c7> lazy-loaded attributes: resources,numa_topology {{(pid=61947) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1146.762268] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948af055-3018-4567-9f8a-0c842444125f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.771114] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31452a8c-58f4-4c79-82d5-94df156ca615 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.801494] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4056ca-3750-46fe-8a24-ee55b5c20d36 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.808324] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f3bdec-d6d6-4cc3-bd90-82ef02593c97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.820788] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 683d40c9-b738-486a-806a-7e895637ee85] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1146.822769] env[61947]: DEBUG nova.compute.provider_tree [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.009923] env[61947]: DEBUG nova.network.neutron [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updated VIF entry in instance network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1147.010342] env[61947]: DEBUG nova.network.neutron [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": null, "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5c342e49-9c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.326200] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: 607d29ad-2db6-4146-a0df-192f727e9d31] Instance has had 0 of 5 cleanup attempts {{(pid=61947) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1147.328650] env[61947]: DEBUG nova.scheduler.client.report [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1147.331483] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.512729] env[61947]: DEBUG oslo_concurrency.lockutils [req-185ff9c4-70ac-4b46-8a6c-d0b11871121c req-ed04ccf7-3ee5-4a3c-ac2f-0bbadfb96d5c service nova] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.834043] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.114s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.836386] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.836537] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Cleaning up deleted instances with incomplete migration {{(pid=61947) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1148.342925] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1148.343559] env[61947]: DEBUG oslo_concurrency.lockutils [None req-1866227a-7060-4d9b-af8a-9a397a32f07e tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.816s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.344351] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.013s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.344536] env[61947]: INFO nova.compute.manager [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Unshelving [ 1149.344959] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.367165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.367441] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.367650] env[61947]: DEBUG nova.objects.instance [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'pci_requests' on Instance uuid 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.556720] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Volume attach. Driver type: vmdk {{(pid=61947) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1149.556976] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264748', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'name': 'volume-54893030-1423-4762-aec4-3e6074904507', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9a31c8d-da69-45a6-9921-3a587d212f3d', 'attached_at': '', 'detached_at': '', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'serial': '54893030-1423-4762-aec4-3e6074904507'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1149.557889] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1769dfd8-fcfe-420b-b6e0-3bf3d1838918 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.578388] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2024a61a-7444-4b4e-b941-0a579d1e9df0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.601730] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] volume-54893030-1423-4762-aec4-3e6074904507/volume-54893030-1423-4762-aec4-3e6074904507.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.601993] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf31da4b-71ea-48e3-aea5-cae7eda98ad1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.619036] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1149.619036] env[61947]: value = "task-1225043" [ 1149.619036] env[61947]: _type = "Task" [ 1149.619036] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.626732] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.849883] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.850093] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.850333] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.850426] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.871924] env[61947]: DEBUG nova.objects.instance [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'numa_topology' on Instance uuid 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.128457] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225043, 'name': ReconfigVM_Task, 'duration_secs': 0.350138} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.128696] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfigured VM instance instance-00000066 to attach disk [datastore2] volume-54893030-1423-4762-aec4-3e6074904507/volume-54893030-1423-4762-aec4-3e6074904507.vmdk or device None with type thin {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.133336] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-373c864a-e024-470a-ad4c-d2a22d0e49bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.148430] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1150.148430] env[61947]: value = "task-1225044" [ 1150.148430] env[61947]: _type = "Task" [ 1150.148430] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.155752] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.354118] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Getting list of instances from cluster (obj){ [ 1150.354118] env[61947]: value = "domain-c8" [ 1150.354118] env[61947]: _type = "ClusterComputeResource" [ 1150.354118] env[61947]: } {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1150.355208] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77819905-5df3-4239-a6cb-bc925ad01f1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.367340] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Got total of 1 instances {{(pid=61947) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1150.367509] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Triggering sync for uuid b9a31c8d-da69-45a6-9921-3a587d212f3d {{(pid=61947) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1150.367854] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.368068] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.368202] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1150.373758] env[61947]: INFO nova.compute.claims [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1150.658255] env[61947]: DEBUG oslo_vmware.api [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225044, 'name': ReconfigVM_Task, 'duration_secs': 0.125759} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.658583] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264748', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'name': 'volume-54893030-1423-4762-aec4-3e6074904507', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9a31c8d-da69-45a6-9921-3a587d212f3d', 'attached_at': '', 'detached_at': '', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'serial': '54893030-1423-4762-aec4-3e6074904507'} {{(pid=61947) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1151.417944] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa8557-768f-439b-8f72-7b27edfb34ce {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.425722] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2789c2a8-620c-43cd-8b08-503d7cf8b0a1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.455405] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317bcfc2-011c-4177-9814-e7bfcb3d4d39 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.462320] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a626072-5445-4911-a5c8-f21606ef07b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.475306] env[61947]: DEBUG nova.compute.provider_tree [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.693917] env[61947]: DEBUG nova.objects.instance [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid b9a31c8d-da69-45a6-9921-3a587d212f3d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.978244] env[61947]: DEBUG nova.scheduler.client.report [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.200072] env[61947]: DEBUG oslo_concurrency.lockutils [None req-8505ef6e-5f34-428c-a464-ac08ee7915c7 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.201027] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.833s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.202084] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987a70b2-81af-4a55-b445-05867c828ac8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.376572] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.483386] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.116s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.511865] env[61947]: INFO nova.network.neutron [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating port 5c342e49-9cb5-4596-ae93-6dd705d98225 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1152.711050] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.711304] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.335s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.217871] env[61947]: INFO nova.compute.manager [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Detaching volume 54893030-1423-4762-aec4-3e6074904507 [ 1153.249411] env[61947]: INFO nova.virt.block_device [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Attempting to driver detach volume 54893030-1423-4762-aec4-3e6074904507 from mountpoint /dev/sdb [ 1153.249651] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Volume detach. Driver type: vmdk {{(pid=61947) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1153.249837] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264748', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'name': 'volume-54893030-1423-4762-aec4-3e6074904507', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9a31c8d-da69-45a6-9921-3a587d212f3d', 'attached_at': '', 'detached_at': '', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'serial': '54893030-1423-4762-aec4-3e6074904507'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1153.250748] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff27f579-8a64-4990-a499-0baf2ab68163 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.271481] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d12fce3-c683-40c2-9246-5a8757790b0e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.278279] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648dfac2-2242-4bf1-85fe-4dd0382e71c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.297550] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b36d83-f3af-4c43-8bb3-c14fadbc400f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.313007] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] The volume has not been displaced from its original location: [datastore2] volume-54893030-1423-4762-aec4-3e6074904507/volume-54893030-1423-4762-aec4-3e6074904507.vmdk. No consolidation needed. {{(pid=61947) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1153.318079] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1153.318356] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11a15fe6-3bcd-4400-a822-82a26e4bfec1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.335118] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1153.335118] env[61947]: value = "task-1225046" [ 1153.335118] env[61947]: _type = "Task" [ 1153.335118] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.342744] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225046, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.844662] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225046, 'name': ReconfigVM_Task, 'duration_secs': 0.205655} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.844941] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=61947) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1153.849464] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad67d25d-2267-4af4-945d-1fc75364b6d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.864240] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1153.864240] env[61947]: value = "task-1225047" [ 1153.864240] env[61947]: _type = "Task" [ 1153.864240] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.871964] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.901127] env[61947]: DEBUG nova.compute.manager [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.901436] env[61947]: DEBUG oslo_concurrency.lockutils [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.901701] env[61947]: DEBUG oslo_concurrency.lockutils [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.901908] env[61947]: DEBUG oslo_concurrency.lockutils [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.902144] env[61947]: DEBUG nova.compute.manager [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] No waiting events found dispatching network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1153.902328] env[61947]: WARNING nova.compute.manager [req-f4f8245c-6d25-4217-b3c0-083a8947f776 req-c80beaea-be0d-4312-882f-6ada39396ea9 service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received unexpected event network-vif-plugged-5c342e49-9cb5-4596-ae93-6dd705d98225 for instance with vm_state shelved_offloaded and task_state spawning. [ 1153.986979] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.987211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.987397] env[61947]: DEBUG nova.network.neutron [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1154.374807] env[61947]: DEBUG oslo_vmware.api [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225047, 'name': ReconfigVM_Task, 'duration_secs': 0.124996} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.375127] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264748', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'name': 'volume-54893030-1423-4762-aec4-3e6074904507', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9a31c8d-da69-45a6-9921-3a587d212f3d', 'attached_at': '', 'detached_at': '', 'volume_id': '54893030-1423-4762-aec4-3e6074904507', 'serial': '54893030-1423-4762-aec4-3e6074904507'} {{(pid=61947) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1154.707746] env[61947]: DEBUG nova.network.neutron [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.917344] env[61947]: DEBUG nova.objects.instance [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'flavor' on Instance uuid b9a31c8d-da69-45a6-9921-3a587d212f3d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.210353] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.240924] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1c8f825b399c83031adabb2a24b88eec',container_format='bare',created_at=2024-10-09T14:54:34Z,direct_url=,disk_format='vmdk',id=f02d075c-e79e-4232-8cda-baac34044bf4,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1515261263-shelved',owner='5273b44dfca848538176692ab2a19e3d',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-10-09T14:54:47Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1155.241225] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1155.241415] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1155.241621] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1155.241781] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1155.241946] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1155.242189] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1155.242355] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1155.242527] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1155.242734] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1155.242922] env[61947]: DEBUG nova.virt.hardware [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.243833] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61867674-2ba8-483c-b2a3-1b320ecd3a6c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.252052] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca15898f-9e22-4693-8011-7b4484a7e19e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.267136] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:43:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c342e49-9cb5-4596-ae93-6dd705d98225', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1155.272841] env[61947]: DEBUG oslo.service.loopingcall [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.273440] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1155.273440] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2a36f24-ed8e-42f2-9f4d-7e6eedc5f8b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.292922] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1155.292922] env[61947]: value = "task-1225048" [ 1155.292922] env[61947]: _type = "Task" [ 1155.292922] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.300486] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225048, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.803448] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225048, 'name': CreateVM_Task, 'duration_secs': 0.315616} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.803815] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1155.805300] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.805524] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.806080] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1155.806406] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44af8b74-21f0-44fc-b3e9-43351ed875d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.811681] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1155.811681] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52539567-428f-0790-0717-00bc76632730" [ 1155.811681] env[61947]: _type = "Task" [ 1155.811681] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.823756] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52539567-428f-0790-0717-00bc76632730, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.925368] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e3f2ded0-f3af-4c04-89c0-f72052b45810 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.214s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.938666] env[61947]: DEBUG nova.compute.manager [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1155.938881] env[61947]: DEBUG nova.compute.manager [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing instance network info cache due to event network-changed-5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1155.939114] env[61947]: DEBUG oslo_concurrency.lockutils [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] Acquiring lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.939312] env[61947]: DEBUG oslo_concurrency.lockutils [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] Acquired lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.939416] env[61947]: DEBUG nova.network.neutron [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Refreshing network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1156.322213] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.322490] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Processing image f02d075c-e79e-4232-8cda-baac34044bf4 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.322779] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.322944] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.323143] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1156.323390] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4460c137-7e10-41f4-8e84-d67ff6d90184 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.330998] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1156.331190] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1156.331881] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d870765-126b-442a-9c63-af06939734c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.336447] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1156.336447] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523a51a3-52e7-7219-9d71-708fa83bac3c" [ 1156.336447] env[61947]: _type = "Task" [ 1156.336447] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.343791] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]523a51a3-52e7-7219-9d71-708fa83bac3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.644539] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.644804] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.645035] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.645223] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.645391] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.647731] env[61947]: INFO nova.compute.manager [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Terminating instance [ 1156.649438] env[61947]: DEBUG nova.compute.manager [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1156.649639] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1156.650450] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56daad2-4cca-4929-8dd5-b57f8b94309f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.658179] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1156.658408] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8d5f527-022a-45f9-9054-837295ce5348 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.663897] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1156.663897] env[61947]: value = "task-1225049" [ 1156.663897] env[61947]: _type = "Task" [ 1156.663897] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.671502] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.678014] env[61947]: DEBUG nova.network.neutron [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updated VIF entry in instance network info cache for port 5c342e49-9cb5-4596-ae93-6dd705d98225. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1156.678352] env[61947]: DEBUG nova.network.neutron [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [{"id": "5c342e49-9cb5-4596-ae93-6dd705d98225", "address": "fa:16:3e:7b:43:0f", "network": {"id": "b52842a0-6442-4f29-9aa1-1e61a06487bd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-193104640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5273b44dfca848538176692ab2a19e3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c342e49-9c", "ovs_interfaceid": "5c342e49-9cb5-4596-ae93-6dd705d98225", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.847602] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Preparing fetch location {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1156.847900] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Fetch image to [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7/OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7.vmdk {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1156.848111] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Downloading stream optimized image f02d075c-e79e-4232-8cda-baac34044bf4 to [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7/OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7.vmdk on the data store datastore2 as vApp {{(pid=61947) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1156.848291] env[61947]: DEBUG nova.virt.vmwareapi.images [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Downloading image file data f02d075c-e79e-4232-8cda-baac34044bf4 to the ESX as VM named 'OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7' {{(pid=61947) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1156.919279] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1156.919279] env[61947]: value = "resgroup-9" [ 1156.919279] env[61947]: _type = "ResourcePool" [ 1156.919279] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1156.919582] env[61947]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5e5f40ad-e2a2-4bd8-96f7-e34b6b65b65e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.939772] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease: (returnval){ [ 1156.939772] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1156.939772] env[61947]: _type = "HttpNfcLease" [ 1156.939772] env[61947]: } obtained for vApp import into resource pool (val){ [ 1156.939772] env[61947]: value = "resgroup-9" [ 1156.939772] env[61947]: _type = "ResourcePool" [ 1156.939772] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1156.940305] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the lease: (returnval){ [ 1156.940305] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1156.940305] env[61947]: _type = "HttpNfcLease" [ 1156.940305] env[61947]: } to be ready. {{(pid=61947) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1156.946432] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1156.946432] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1156.946432] env[61947]: _type = "HttpNfcLease" [ 1156.946432] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1157.174915] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225049, 'name': PowerOffVM_Task, 'duration_secs': 0.18303} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.175201] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1157.175376] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1157.175632] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3fee4c5-23b9-4e44-aa3f-4bedfcf98b4f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.180849] env[61947]: DEBUG oslo_concurrency.lockutils [req-ce3d6c04-a7ad-4547-8a81-77185dd60c22 req-6a387124-5549-41d5-9669-4685559abede service nova] Releasing lock "refresh_cache-41dc3d63-d455-4f74-9ec8-a7806f25b2c7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.245719] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1157.245935] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1157.246135] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleting the datastore file [datastore1] b9a31c8d-da69-45a6-9921-3a587d212f3d {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1157.246409] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-977fb111-fc1d-4f25-a31d-08c82f6bdb81 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.252054] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for the task: (returnval){ [ 1157.252054] env[61947]: value = "task-1225052" [ 1157.252054] env[61947]: _type = "Task" [ 1157.252054] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.261664] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.447862] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1157.447862] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1157.447862] env[61947]: _type = "HttpNfcLease" [ 1157.447862] env[61947]: } is initializing. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1157.763248] env[61947]: DEBUG oslo_vmware.api [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Task: {'id': task-1225052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128379} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.764032] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1157.764554] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1157.764742] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1157.765031] env[61947]: INFO nova.compute.manager [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1157.765743] env[61947]: DEBUG oslo.service.loopingcall [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1157.765947] env[61947]: DEBUG nova.compute.manager [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1157.767163] env[61947]: DEBUG nova.network.neutron [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1157.950142] env[61947]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1157.950142] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1157.950142] env[61947]: _type = "HttpNfcLease" [ 1157.950142] env[61947]: } is ready. {{(pid=61947) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1157.950625] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1157.950625] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ebc793-19e2-6283-f243-2eea958cc365" [ 1157.950625] env[61947]: _type = "HttpNfcLease" [ 1157.950625] env[61947]: }. {{(pid=61947) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1157.951138] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea8572c-fb58-48f1-83f1-b85ae6f548fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.958284] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk from lease info. {{(pid=61947) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1157.958461] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk. {{(pid=61947) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1158.022620] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6d30e1b8-8ce1-4048-a90e-cb14a27b288b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.208204] env[61947]: DEBUG nova.compute.manager [req-fe37977e-8465-4c78-8b2d-32311a5c3036 req-539bed99-ead0-4a3f-82c4-a45a674cfec5 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Received event network-vif-deleted-746ecfed-7986-4f58-87a1-1bdddd80d0b2 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1158.208398] env[61947]: INFO nova.compute.manager [req-fe37977e-8465-4c78-8b2d-32311a5c3036 req-539bed99-ead0-4a3f-82c4-a45a674cfec5 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Neutron deleted interface 746ecfed-7986-4f58-87a1-1bdddd80d0b2; detaching it from the instance and deleting it from the info cache [ 1158.208580] env[61947]: DEBUG nova.network.neutron [req-fe37977e-8465-4c78-8b2d-32311a5c3036 req-539bed99-ead0-4a3f-82c4-a45a674cfec5 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.680800] env[61947]: DEBUG nova.network.neutron [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.711430] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58ac6e01-f951-4656-9a73-e1e2ecc046f4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.723189] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2fb45c-c916-4bd6-9e07-d46030d09494 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.749974] env[61947]: DEBUG nova.compute.manager [req-fe37977e-8465-4c78-8b2d-32311a5c3036 req-539bed99-ead0-4a3f-82c4-a45a674cfec5 service nova] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Detach interface failed, port_id=746ecfed-7986-4f58-87a1-1bdddd80d0b2, reason: Instance b9a31c8d-da69-45a6-9921-3a587d212f3d could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1159.099406] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Completed reading data from the image iterator. {{(pid=61947) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1159.099787] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1159.100865] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d78508-a2a8-4b1b-aac0-ff33468e6c18 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.107269] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk is in state: ready. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1159.107430] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk. {{(pid=61947) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1159.107664] env[61947]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-334908ed-7d00-4ac4-aa47-d0fe66b693d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.184055] env[61947]: INFO nova.compute.manager [-] [instance: b9a31c8d-da69-45a6-9921-3a587d212f3d] Took 1.42 seconds to deallocate network for instance. [ 1159.303289] env[61947]: DEBUG oslo_vmware.rw_handles [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e59c5c-ef23-ff9f-7707-6666a51abaa0/disk-0.vmdk. {{(pid=61947) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1159.303598] env[61947]: INFO nova.virt.vmwareapi.images [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Downloaded image file data f02d075c-e79e-4232-8cda-baac34044bf4 [ 1159.304376] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4a442e-3f21-41f4-bfc2-de80aea62084 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.319081] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73f44793-c714-4ed1-bc1a-8b6eb01a2de5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.345065] env[61947]: INFO nova.virt.vmwareapi.images [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] The imported VM was unregistered [ 1159.347741] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Caching image {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1159.347971] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Creating directory with path [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1159.348550] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94672dde-aca8-4f28-99b0-0269d4d25c31 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.358206] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Created directory with path [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4 {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1159.358385] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7/OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7.vmdk to [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk. {{(pid=61947) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1159.358617] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4e020d93-37f5-401b-9faf-e2beaf2b79bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.365508] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1159.365508] env[61947]: value = "task-1225054" [ 1159.365508] env[61947]: _type = "Task" [ 1159.365508] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.372432] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.691538] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.691962] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.692237] env[61947]: DEBUG nova.objects.instance [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lazy-loading 'resources' on Instance uuid b9a31c8d-da69-45a6-9921-3a587d212f3d {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.875274] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.242214] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c533050-81c7-4d24-bdd5-bd95a7dc89bf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.250588] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c67643-c9fb-458a-9526-48c70d73d60a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.283520] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1563dee-ab60-46cf-8ca0-62556118a4ca {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.292372] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dbaa9b-3c2b-4db9-8773-0b45540e6e0c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.306746] env[61947]: DEBUG nova.compute.provider_tree [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.376948] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.810675] env[61947]: DEBUG nova.scheduler.client.report [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1160.876597] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.315633] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.334518] env[61947]: INFO nova.scheduler.client.report [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Deleted allocations for instance b9a31c8d-da69-45a6-9921-3a587d212f3d [ 1161.380057] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.843525] env[61947]: DEBUG oslo_concurrency.lockutils [None req-19ce5581-5b25-4e44-a683-270212197079 tempest-AttachVolumeNegativeTest-1917565154 tempest-AttachVolumeNegativeTest-1917565154-project-member] Lock "b9a31c8d-da69-45a6-9921-3a587d212f3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.198s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.880227] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225054, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.232865} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.880227] env[61947]: INFO nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7/OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7.vmdk to [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk. [ 1161.880227] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Cleaning up location [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1161.880227] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e15c8460-3fa9-4eea-a0d2-ccfa706bd5f7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.880227] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb06810c-05c7-48e2-b3a6-cc335ffb2b53 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.884799] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1161.884799] env[61947]: value = "task-1225055" [ 1161.884799] env[61947]: _type = "Task" [ 1161.884799] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.891861] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.393655] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091753} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.394053] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.394115] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.394319] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk to [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1162.394562] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2899037f-0c30-45a0-9e7a-194749774d59 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.400778] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1162.400778] env[61947]: value = "task-1225057" [ 1162.400778] env[61947]: _type = "Task" [ 1162.400778] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.407921] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.914820] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.411847] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.912696] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.414397] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.914460] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225057, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.105438} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.914713] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f02d075c-e79e-4232-8cda-baac34044bf4/f02d075c-e79e-4232-8cda-baac34044bf4.vmdk to [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1164.915503] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c150862-2c84-4203-b5d3-1408cb3aa4ac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.945774] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.946132] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c02cb197-42b1-4e8d-b05e-af53dabe7899 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.966216] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1164.966216] env[61947]: value = "task-1225058" [ 1164.966216] env[61947]: _type = "Task" [ 1164.966216] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.974048] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225058, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.475775] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225058, 'name': ReconfigVM_Task, 'duration_secs': 0.445642} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.476152] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7/41dc3d63-d455-4f74-9ec8-a7806f25b2c7.vmdk or device None with type streamOptimized {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.476642] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-183ae476-4460-47be-a70e-17c4810d6fe4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.482338] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1165.482338] env[61947]: value = "task-1225060" [ 1165.482338] env[61947]: _type = "Task" [ 1165.482338] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.490923] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225060, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.992361] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225060, 'name': Rename_Task, 'duration_secs': 0.234441} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.992644] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1165.992919] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09a70e67-ad47-48ad-88d6-cbbc662502a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.998993] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1165.998993] env[61947]: value = "task-1225061" [ 1165.998993] env[61947]: _type = "Task" [ 1165.998993] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.005988] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.508851] env[61947]: DEBUG oslo_vmware.api [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225061, 'name': PowerOnVM_Task, 'duration_secs': 0.446149} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.509221] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1166.684110] env[61947]: DEBUG nova.compute.manager [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.685034] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175d203f-4c44-471e-b8cf-afc50e1ab3cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.203593] env[61947]: DEBUG oslo_concurrency.lockutils [None req-11add39b-06a7-4b85-b6b0-47dce1d933d0 tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.859s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.495085] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.495085] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.495503] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.495503] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.495607] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.498019] env[61947]: INFO nova.compute.manager [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Terminating instance [ 1168.499953] env[61947]: DEBUG nova.compute.manager [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1168.500165] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1168.500999] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5175930-0a38-4e0e-aad3-fb0a44f16f20 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.507953] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1168.508191] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29730a4f-32ae-412f-bdf2-73dfc2bc1130 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.513811] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1168.513811] env[61947]: value = "task-1225063" [ 1168.513811] env[61947]: _type = "Task" [ 1168.513811] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.521183] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.025157] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225063, 'name': PowerOffVM_Task, 'duration_secs': 0.154582} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.025476] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1169.025676] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1169.025939] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55052215-4ac0-47f2-8269-386375b688fe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.084186] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1169.084416] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1169.084655] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleting the datastore file [datastore2] 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.084938] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-016bfc09-8535-4c46-90d7-c139d93c5960 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.090910] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for the task: (returnval){ [ 1169.090910] env[61947]: value = "task-1225065" [ 1169.090910] env[61947]: _type = "Task" [ 1169.090910] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.098309] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.600815] env[61947]: DEBUG oslo_vmware.api [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Task: {'id': task-1225065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119982} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.603183] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1169.603183] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1169.603183] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1169.603183] env[61947]: INFO nova.compute.manager [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1169.603183] env[61947]: DEBUG oslo.service.loopingcall [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.603183] env[61947]: DEBUG nova.compute.manager [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1169.603183] env[61947]: DEBUG nova.network.neutron [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1170.016494] env[61947]: DEBUG nova.compute.manager [req-a7dba5f0-ba56-49f0-8392-74922dc5932b req-9cd68c40-ab67-4cde-bd04-6335bd5e829e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Received event network-vif-deleted-5c342e49-9cb5-4596-ae93-6dd705d98225 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1170.016494] env[61947]: INFO nova.compute.manager [req-a7dba5f0-ba56-49f0-8392-74922dc5932b req-9cd68c40-ab67-4cde-bd04-6335bd5e829e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Neutron deleted interface 5c342e49-9cb5-4596-ae93-6dd705d98225; detaching it from the instance and deleting it from the info cache [ 1170.016494] env[61947]: DEBUG nova.network.neutron [req-a7dba5f0-ba56-49f0-8392-74922dc5932b req-9cd68c40-ab67-4cde-bd04-6335bd5e829e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.492402] env[61947]: DEBUG nova.network.neutron [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.518653] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f898fd6-f3cd-46ae-bf92-63a2c06f6eee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.529035] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec60ee6-8575-4fff-83a4-1a4cf12dcc0d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.550818] env[61947]: DEBUG nova.compute.manager [req-a7dba5f0-ba56-49f0-8392-74922dc5932b req-9cd68c40-ab67-4cde-bd04-6335bd5e829e service nova] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Detach interface failed, port_id=5c342e49-9cb5-4596-ae93-6dd705d98225, reason: Instance 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1170.995377] env[61947]: INFO nova.compute.manager [-] [instance: 41dc3d63-d455-4f74-9ec8-a7806f25b2c7] Took 1.39 seconds to deallocate network for instance. [ 1171.501943] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.502383] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.502688] env[61947]: DEBUG nova.objects.instance [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lazy-loading 'resources' on Instance uuid 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.036562] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d95a7d-331c-488f-8d09-13a2456f78e8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.044066] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757c91af-4a1f-4ca5-be9b-5f0fb6a29d74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.072785] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28397187-0418-43af-9f9a-8d2c7d9beb00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.079418] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f754c80d-b99b-4174-9e48-053cd2e1c4fa {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.093190] env[61947]: DEBUG nova.compute.provider_tree [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.596220] env[61947]: DEBUG nova.scheduler.client.report [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1173.100671] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.122728] env[61947]: INFO nova.scheduler.client.report [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Deleted allocations for instance 41dc3d63-d455-4f74-9ec8-a7806f25b2c7 [ 1173.634670] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c48a2da5-9e32-4ae3-b487-76effcf4b81f tempest-ServerActionsTestOtherB-1107084526 tempest-ServerActionsTestOtherB-1107084526-project-member] Lock "41dc3d63-d455-4f74-9ec8-a7806f25b2c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.140s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.904473] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1174.904831] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1174.904831] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1175.408357] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Didn't find any instances for network info cache update. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1176.386163] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1176.890279] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.891023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.891023] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.891023] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1176.892129] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954fe5c2-4805-4b62-b1e9-253cdefd13d3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.902082] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c080f23-e3e9-4153-ad41-43cdfc071673 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.916506] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f650f60-b638-48dc-a896-704fc5a4f620 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.922722] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee86298-ec1e-42bf-af45-282ab04f1448 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.966924] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180739MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1176.967136] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.967343] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.986643] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1177.986884] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1177.999729] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba449ad4-bc44-4ef6-91b4-29771e26be84 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.007872] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990e2ac0-26b4-458c-9eb8-0531ab5119a9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.039079] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9bfddd-55bf-496e-a06f-8c405a92e8db {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.046823] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dc3f72-37b1-4ae9-b841-2b3bc59c2a2b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.059638] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.562103] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.068814] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1179.069118] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.102s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.065052] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.065579] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.065769] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.065921] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.066083] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.066235] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.066523] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1181.386650] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.887995] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.888290] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.391493] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1182.912866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.913183] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.914687] env[61947]: INFO nova.compute.claims [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1183.952637] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d9d014-5463-47d5-be6b-05c6b510cfb2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.959727] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f9ec7a-f8c7-46f5-9997-c22fa69d2069 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.989106] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2141690-85ff-4383-ab50-0811e15be0c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.995496] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071344a9-4bba-473c-8579-710812a3ae9f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.007687] env[61947]: DEBUG nova.compute.provider_tree [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.511053] env[61947]: DEBUG nova.scheduler.client.report [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1185.015592] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.016160] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1185.520891] env[61947]: DEBUG nova.compute.utils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1185.522591] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1185.522762] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1185.561664] env[61947]: DEBUG nova.policy [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1185.801821] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Successfully created port: 16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1186.026057] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1187.034662] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1187.059241] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1187.059499] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1187.059661] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.059846] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1187.059996] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.060162] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1187.060371] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1187.060533] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1187.060701] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1187.060862] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1187.061049] env[61947]: DEBUG nova.virt.hardware [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1187.061952] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe6c548-1747-491e-a8f8-872a2a006a3a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.069682] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c400fe8-dfee-423e-8e29-433873cd403d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.214301] env[61947]: DEBUG nova.compute.manager [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Received event network-vif-plugged-16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1187.214496] env[61947]: DEBUG oslo_concurrency.lockutils [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] Acquiring lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.214715] env[61947]: DEBUG oslo_concurrency.lockutils [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.214888] env[61947]: DEBUG oslo_concurrency.lockutils [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.215073] env[61947]: DEBUG nova.compute.manager [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] No waiting events found dispatching network-vif-plugged-16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1187.215248] env[61947]: WARNING nova.compute.manager [req-03fbb91d-185f-4a32-bf3e-6ea27c3a17c9 req-948f485c-f12f-4f90-9790-9511ecf3fa40 service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Received unexpected event network-vif-plugged-16b0fd1c-1dae-455c-ba72-c515ff774b9a for instance with vm_state building and task_state spawning. [ 1187.267766] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Successfully updated port: 16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1187.770447] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.770605] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.770761] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1188.302065] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1188.422725] env[61947]: DEBUG nova.network.neutron [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [{"id": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "address": "fa:16:3e:b9:f2:a9", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16b0fd1c-1d", "ovs_interfaceid": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.925860] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.926255] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Instance network_info: |[{"id": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "address": "fa:16:3e:b9:f2:a9", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16b0fd1c-1d", "ovs_interfaceid": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1188.926753] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:f2:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16b0fd1c-1dae-455c-ba72-c515ff774b9a', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1188.934429] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating folder: Project (38ef945423f04aa6b5e50d13138d08a0). Parent ref: group-v264556. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1188.934718] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffe020c0-963e-4d0d-9752-1c1c5ea15688 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.945584] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created folder: Project (38ef945423f04aa6b5e50d13138d08a0) in parent group-v264556. [ 1188.945776] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating folder: Instances. Parent ref: group-v264751. {{(pid=61947) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1188.945997] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aeb30bf9-084e-45bf-8fbc-6891b5b09800 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.953643] env[61947]: INFO nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created folder: Instances in parent group-v264751. [ 1188.953858] env[61947]: DEBUG oslo.service.loopingcall [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1188.954051] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1188.954250] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40f56ebd-02d4-450c-b91f-a44d296b810e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.971720] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1188.971720] env[61947]: value = "task-1225069" [ 1188.971720] env[61947]: _type = "Task" [ 1188.971720] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.978551] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225069, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.239995] env[61947]: DEBUG nova.compute.manager [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Received event network-changed-16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.240231] env[61947]: DEBUG nova.compute.manager [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Refreshing instance network info cache due to event network-changed-16b0fd1c-1dae-455c-ba72-c515ff774b9a. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1189.240453] env[61947]: DEBUG oslo_concurrency.lockutils [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] Acquiring lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.240604] env[61947]: DEBUG oslo_concurrency.lockutils [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] Acquired lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.240770] env[61947]: DEBUG nova.network.neutron [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Refreshing network info cache for port 16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1189.480998] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225069, 'name': CreateVM_Task, 'duration_secs': 0.276455} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.481341] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1189.481804] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.481973] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.482317] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1189.482570] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-463985e3-f9a8-4abb-8ef4-d574e03c1d7b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.486808] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1189.486808] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52da780d-3d45-4639-74a1-ab57397409a0" [ 1189.486808] env[61947]: _type = "Task" [ 1189.486808] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.493838] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52da780d-3d45-4639-74a1-ab57397409a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.914102] env[61947]: DEBUG nova.network.neutron [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updated VIF entry in instance network info cache for port 16b0fd1c-1dae-455c-ba72-c515ff774b9a. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1189.914470] env[61947]: DEBUG nova.network.neutron [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [{"id": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "address": "fa:16:3e:b9:f2:a9", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16b0fd1c-1d", "ovs_interfaceid": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.997060] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52da780d-3d45-4639-74a1-ab57397409a0, 'name': SearchDatastore_Task, 'duration_secs': 0.00955} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.997362] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.997613] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1189.997855] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.998013] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.998201] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1189.998453] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7527d5c8-e953-4036-aef8-d39438365a22 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.005941] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1190.006176] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1190.006844] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29473625-3ed8-4941-bb7b-b58972b6109d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.011413] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1190.011413] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b87e25-8394-aad6-c3dd-afa7c7f0cbec" [ 1190.011413] env[61947]: _type = "Task" [ 1190.011413] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.018438] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b87e25-8394-aad6-c3dd-afa7c7f0cbec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.417555] env[61947]: DEBUG oslo_concurrency.lockutils [req-18eca6d9-7c04-4038-a76e-e989962e110f req-57940132-9ee3-4e58-96bc-f6dd1748172e service nova] Releasing lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.521930] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52b87e25-8394-aad6-c3dd-afa7c7f0cbec, 'name': SearchDatastore_Task, 'duration_secs': 0.00864} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.522703] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c12c620d-06cd-4874-b6a4-1e45e6b4b88a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.527462] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1190.527462] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b9a1a-c899-3c2f-c160-2e337fc1d6be" [ 1190.527462] env[61947]: _type = "Task" [ 1190.527462] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.535709] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b9a1a-c899-3c2f-c160-2e337fc1d6be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.037669] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521b9a1a-c899-3c2f-c160-2e337fc1d6be, 'name': SearchDatastore_Task, 'duration_secs': 0.010064} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.038092] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.038198] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8/ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1191.038459] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-244c448b-0ca9-457a-909f-f15dcf8f5018 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.044675] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1191.044675] env[61947]: value = "task-1225070" [ 1191.044675] env[61947]: _type = "Task" [ 1191.044675] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.052080] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.554402] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.403977} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.554849] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore1] ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8/ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1191.554849] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1191.555115] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e50898dd-c271-4930-9277-abc926c0addc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.561275] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1191.561275] env[61947]: value = "task-1225071" [ 1191.561275] env[61947]: _type = "Task" [ 1191.561275] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.568764] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.070391] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225071, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059848} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.070609] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1192.071376] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb652b14-e65a-411f-a13b-5a7c4505588f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.092056] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8/ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.092289] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-907949d6-dd3e-4771-87b8-52115b835965 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.111443] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1192.111443] env[61947]: value = "task-1225072" [ 1192.111443] env[61947]: _type = "Task" [ 1192.111443] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.119778] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.621294] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225072, 'name': ReconfigVM_Task, 'duration_secs': 0.269796} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.621660] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Reconfigured VM instance instance-00000067 to attach disk [datastore1] ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8/ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1192.622222] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-138553d5-3d57-4f68-87f6-09aab00cbed8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.627897] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1192.627897] env[61947]: value = "task-1225073" [ 1192.627897] env[61947]: _type = "Task" [ 1192.627897] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.634774] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225073, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.137682] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225073, 'name': Rename_Task, 'duration_secs': 0.12782} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.137950] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1193.138213] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0fdd766-7bb3-4863-875d-80caefe00958 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.143683] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1193.143683] env[61947]: value = "task-1225074" [ 1193.143683] env[61947]: _type = "Task" [ 1193.143683] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.150558] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225074, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.653426] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225074, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.153983] env[61947]: DEBUG oslo_vmware.api [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225074, 'name': PowerOnVM_Task, 'duration_secs': 0.655921} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.155245] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1194.155245] env[61947]: INFO nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Took 7.12 seconds to spawn the instance on the hypervisor. [ 1194.155245] env[61947]: DEBUG nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1194.155744] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cedce44-0f73-440c-9f27-9efea9c61d0d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.673053] env[61947]: INFO nova.compute.manager [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Took 11.78 seconds to build instance. [ 1195.175993] env[61947]: DEBUG oslo_concurrency.lockutils [None req-14b7aad2-6107-43d5-b460-b7c4232d108a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.288s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.755501] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.755810] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.258598] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1196.778038] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.778311] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.779771] env[61947]: INFO nova.compute.claims [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1197.820578] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39bd041-3601-4910-87b4-8f926909e29b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.828277] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec1a1b8-483c-4cb1-a378-1bc72a1c44d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.858183] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436553e8-fd49-4e5d-84bd-be6ff14c5315 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.864921] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060d9f47-2f14-422b-8561-f094dabd222b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.877464] env[61947]: DEBUG nova.compute.provider_tree [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.380083] env[61947]: DEBUG nova.scheduler.client.report [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.885192] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.107s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.885781] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1199.390941] env[61947]: DEBUG nova.compute.utils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1199.392450] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1199.392627] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1199.448360] env[61947]: DEBUG nova.policy [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1199.688865] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Successfully created port: cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1199.896592] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1200.906334] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1200.932085] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1200.932350] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1200.932512] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1200.932700] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1200.932850] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1200.932998] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1200.933221] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1200.933401] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1200.933592] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1200.933762] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1200.933936] env[61947]: DEBUG nova.virt.hardware [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1200.934830] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51046f84-e713-40e2-bd53-6f25d18ed575 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.942788] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6ed9db-db02-42b1-870b-bb900e3aa3f5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.053053] env[61947]: DEBUG nova.compute.manager [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Received event network-vif-plugged-cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1201.053299] env[61947]: DEBUG oslo_concurrency.lockutils [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] Acquiring lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.053574] env[61947]: DEBUG oslo_concurrency.lockutils [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.053747] env[61947]: DEBUG oslo_concurrency.lockutils [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.053918] env[61947]: DEBUG nova.compute.manager [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] No waiting events found dispatching network-vif-plugged-cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1201.054099] env[61947]: WARNING nova.compute.manager [req-3e97d818-9c0e-4f44-8b47-454cba162428 req-3b49f282-6516-4e53-a477-fa89498a2a43 service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Received unexpected event network-vif-plugged-cad3aa42-c905-49c5-b469-06a7130bf527 for instance with vm_state building and task_state spawning. [ 1201.135105] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Successfully updated port: cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1201.638014] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.638177] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.638321] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1202.170480] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1202.293522] env[61947]: DEBUG nova.network.neutron [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Updating instance_info_cache with network_info: [{"id": "cad3aa42-c905-49c5-b469-06a7130bf527", "address": "fa:16:3e:b4:58:7f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad3aa42-c9", "ovs_interfaceid": "cad3aa42-c905-49c5-b469-06a7130bf527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.795996] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.796304] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Instance network_info: |[{"id": "cad3aa42-c905-49c5-b469-06a7130bf527", "address": "fa:16:3e:b4:58:7f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad3aa42-c9", "ovs_interfaceid": "cad3aa42-c905-49c5-b469-06a7130bf527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1202.796744] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:58:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cad3aa42-c905-49c5-b469-06a7130bf527', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1202.804126] env[61947]: DEBUG oslo.service.loopingcall [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1202.804354] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1202.804583] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2b52f4a-6345-402b-af80-3b5ede6074de {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.824130] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1202.824130] env[61947]: value = "task-1225075" [ 1202.824130] env[61947]: _type = "Task" [ 1202.824130] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.831365] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225075, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.079734] env[61947]: DEBUG nova.compute.manager [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Received event network-changed-cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1203.079969] env[61947]: DEBUG nova.compute.manager [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Refreshing instance network info cache due to event network-changed-cad3aa42-c905-49c5-b469-06a7130bf527. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1203.080383] env[61947]: DEBUG oslo_concurrency.lockutils [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] Acquiring lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1203.080586] env[61947]: DEBUG oslo_concurrency.lockutils [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] Acquired lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.080787] env[61947]: DEBUG nova.network.neutron [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Refreshing network info cache for port cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1203.333975] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225075, 'name': CreateVM_Task, 'duration_secs': 0.27948} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.334296] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1203.334798] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1203.334971] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.335296] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1203.335550] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1854e9e4-bce8-4009-9034-0cca43f71d6f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.339800] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1203.339800] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266ad4b-915d-967b-b1f4-6aeb142fec93" [ 1203.339800] env[61947]: _type = "Task" [ 1203.339800] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.347153] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266ad4b-915d-967b-b1f4-6aeb142fec93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.756184] env[61947]: DEBUG nova.network.neutron [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Updated VIF entry in instance network info cache for port cad3aa42-c905-49c5-b469-06a7130bf527. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1203.756570] env[61947]: DEBUG nova.network.neutron [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Updating instance_info_cache with network_info: [{"id": "cad3aa42-c905-49c5-b469-06a7130bf527", "address": "fa:16:3e:b4:58:7f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad3aa42-c9", "ovs_interfaceid": "cad3aa42-c905-49c5-b469-06a7130bf527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.849910] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5266ad4b-915d-967b-b1f4-6aeb142fec93, 'name': SearchDatastore_Task, 'duration_secs': 0.009371} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.850233] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.850465] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1203.850693] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1203.850844] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.851031] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1203.851285] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2aef95e9-ae8f-4692-8f14-59dbfd790821 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.858908] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1203.859085] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1203.859757] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-251506e1-788b-4402-b1d5-43d5a51b30ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.864319] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1203.864319] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522136d1-ad6a-1cb9-7a95-57eb08511d46" [ 1203.864319] env[61947]: _type = "Task" [ 1203.864319] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.871299] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522136d1-ad6a-1cb9-7a95-57eb08511d46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.259020] env[61947]: DEBUG oslo_concurrency.lockutils [req-ba44cf4d-dc7c-42c9-9195-307ff37ee403 req-9ba631b0-4b35-4db2-86a0-5134eca5068e service nova] Releasing lock "refresh_cache-902d7b0f-55e6-490d-b2dc-4f103fac93b7" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.374733] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]522136d1-ad6a-1cb9-7a95-57eb08511d46, 'name': SearchDatastore_Task, 'duration_secs': 0.007634} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.375497] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b9ca2e4-00ab-4f66-b2a0-0866d1d25313 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.380527] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1204.380527] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a70ebe-0b68-128a-7684-00367bf2acba" [ 1204.380527] env[61947]: _type = "Task" [ 1204.380527] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.388598] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a70ebe-0b68-128a-7684-00367bf2acba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.890989] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a70ebe-0b68-128a-7684-00367bf2acba, 'name': SearchDatastore_Task, 'duration_secs': 0.008464} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.891385] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.891687] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 902d7b0f-55e6-490d-b2dc-4f103fac93b7/902d7b0f-55e6-490d-b2dc-4f103fac93b7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1204.891957] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-957e6f7f-019a-4f14-9be3-68d7962ff5d8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.898719] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1204.898719] env[61947]: value = "task-1225076" [ 1204.898719] env[61947]: _type = "Task" [ 1204.898719] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.906280] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.408787] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.399708} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.409082] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 902d7b0f-55e6-490d-b2dc-4f103fac93b7/902d7b0f-55e6-490d-b2dc-4f103fac93b7.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1205.409269] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1205.409505] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-860ae9f0-4ca5-437f-90e5-7ecab2c845c5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.416352] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1205.416352] env[61947]: value = "task-1225077" [ 1205.416352] env[61947]: _type = "Task" [ 1205.416352] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.423129] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.926541] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055924} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.926818] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1205.927724] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c8ad18-9750-4127-a0f8-f28a439ce797 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.948848] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 902d7b0f-55e6-490d-b2dc-4f103fac93b7/902d7b0f-55e6-490d-b2dc-4f103fac93b7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1205.949088] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e80d42cb-9fcd-472d-a925-db5ebb0bfba0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.967938] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1205.967938] env[61947]: value = "task-1225078" [ 1205.967938] env[61947]: _type = "Task" [ 1205.967938] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.975108] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.478231] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225078, 'name': ReconfigVM_Task, 'duration_secs': 0.278002} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.478696] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 902d7b0f-55e6-490d-b2dc-4f103fac93b7/902d7b0f-55e6-490d-b2dc-4f103fac93b7.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1206.479158] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4f4b4d5-6df3-4f4f-8729-397fe762d453 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.485192] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1206.485192] env[61947]: value = "task-1225079" [ 1206.485192] env[61947]: _type = "Task" [ 1206.485192] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.492837] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225079, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.994624] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225079, 'name': Rename_Task, 'duration_secs': 0.167427} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.994841] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1206.995072] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7860ad2b-e126-4f24-80cb-8cd0417ec42d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.000734] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1207.000734] env[61947]: value = "task-1225080" [ 1207.000734] env[61947]: _type = "Task" [ 1207.000734] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.007995] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.510437] env[61947]: DEBUG oslo_vmware.api [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225080, 'name': PowerOnVM_Task, 'duration_secs': 0.414951} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.510782] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1207.510881] env[61947]: INFO nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Took 6.60 seconds to spawn the instance on the hypervisor. [ 1207.511136] env[61947]: DEBUG nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1207.511917] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db076020-9a60-4ccd-88fc-19e150138532 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.027848] env[61947]: INFO nova.compute.manager [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Took 11.26 seconds to build instance. [ 1208.529534] env[61947]: DEBUG oslo_concurrency.lockutils [None req-44d4606b-aafb-4f42-b190-0298540c84d5 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.774s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.030760] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.031095] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.031322] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.031545] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.031758] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.034093] env[61947]: INFO nova.compute.manager [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Terminating instance [ 1209.035842] env[61947]: DEBUG nova.compute.manager [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1209.036052] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1209.036888] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bd4d0c-055d-49ec-bffc-4b5ed75f8a5d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.045134] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1209.045361] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1afe091-d859-4bdc-a8e8-8a174c4a725c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.051301] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1209.051301] env[61947]: value = "task-1225081" [ 1209.051301] env[61947]: _type = "Task" [ 1209.051301] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.058989] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.562554] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225081, 'name': PowerOffVM_Task, 'duration_secs': 0.197738} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.562947] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1209.563016] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1209.563308] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6a193f6-57e5-47b2-aa1b-dbe2d82dcfa3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.621657] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1209.621888] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1209.622088] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] 902d7b0f-55e6-490d-b2dc-4f103fac93b7 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1209.622359] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-813adc56-b8ee-470e-9470-7d61586782b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.628693] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1209.628693] env[61947]: value = "task-1225083" [ 1209.628693] env[61947]: _type = "Task" [ 1209.628693] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.636471] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.137798] env[61947]: DEBUG oslo_vmware.api [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141774} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.138077] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.138272] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1210.138452] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1210.138629] env[61947]: INFO nova.compute.manager [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1210.138886] env[61947]: DEBUG oslo.service.loopingcall [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.139095] env[61947]: DEBUG nova.compute.manager [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1210.139193] env[61947]: DEBUG nova.network.neutron [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1210.373065] env[61947]: DEBUG nova.compute.manager [req-7ccd5e68-4c42-4e28-8a6c-bc8be28b42f2 req-7db6735f-7828-4df5-a00e-0afefd443e0c service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Received event network-vif-deleted-cad3aa42-c905-49c5-b469-06a7130bf527 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1210.373293] env[61947]: INFO nova.compute.manager [req-7ccd5e68-4c42-4e28-8a6c-bc8be28b42f2 req-7db6735f-7828-4df5-a00e-0afefd443e0c service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Neutron deleted interface cad3aa42-c905-49c5-b469-06a7130bf527; detaching it from the instance and deleting it from the info cache [ 1210.373603] env[61947]: DEBUG nova.network.neutron [req-7ccd5e68-4c42-4e28-8a6c-bc8be28b42f2 req-7db6735f-7828-4df5-a00e-0afefd443e0c service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.846220] env[61947]: DEBUG nova.network.neutron [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.876615] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a7ad20b-6907-4e19-900c-0ba6843e9905 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.887944] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08c0eda-4480-4233-9495-e0a38e8fac74 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.910160] env[61947]: DEBUG nova.compute.manager [req-7ccd5e68-4c42-4e28-8a6c-bc8be28b42f2 req-7db6735f-7828-4df5-a00e-0afefd443e0c service nova] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Detach interface failed, port_id=cad3aa42-c905-49c5-b469-06a7130bf527, reason: Instance 902d7b0f-55e6-490d-b2dc-4f103fac93b7 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1211.348262] env[61947]: INFO nova.compute.manager [-] [instance: 902d7b0f-55e6-490d-b2dc-4f103fac93b7] Took 1.21 seconds to deallocate network for instance. [ 1211.855211] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.855622] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.855941] env[61947]: DEBUG nova.objects.instance [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid 902d7b0f-55e6-490d-b2dc-4f103fac93b7 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1212.398270] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0f66f2-5e9e-4347-92f6-5cb702d389a5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.405796] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36a6fb3-110b-4e08-9a29-3118be0ddc00 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.434957] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee789213-a299-4f72-9045-ef16a4359249 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.442318] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70912ea0-1f11-48f8-8bd9-4541a76b5aff {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.455305] env[61947]: DEBUG nova.compute.provider_tree [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.958445] env[61947]: DEBUG nova.scheduler.client.report [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1213.464939] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.482622] env[61947]: INFO nova.scheduler.client.report [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance 902d7b0f-55e6-490d-b2dc-4f103fac93b7 [ 1213.991376] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e347ac88-c044-447c-a523-a576e07625a0 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "902d7b0f-55e6-490d-b2dc-4f103fac93b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.960s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.920681] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "5bc960ad-3a21-465e-b6b8-83f04064745c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.920928] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.423714] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1215.943485] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.943770] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.945459] env[61947]: INFO nova.compute.claims [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1216.993403] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3510c26e-05e6-4eea-b1d3-5a8d68363d78 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.001111] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253dfe0d-5ff1-45c3-8023-416e2c5b13bd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.030289] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eae836-08b8-4ba9-917f-1509b7ba4fa9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.037044] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa2be72-c51a-4362-bc25-bb4c3d7ea13f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.049504] env[61947]: DEBUG nova.compute.provider_tree [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.552422] env[61947]: DEBUG nova.scheduler.client.report [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1218.058135] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.058692] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1218.563503] env[61947]: DEBUG nova.compute.utils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1218.564944] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1218.565130] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1218.613184] env[61947]: DEBUG nova.policy [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1218.845480] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Successfully created port: 9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1219.068007] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1220.078977] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1220.103648] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1220.103942] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1220.104130] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1220.104324] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1220.104475] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1220.104623] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1220.104872] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1220.105167] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1220.105253] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1220.105387] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1220.105560] env[61947]: DEBUG nova.virt.hardware [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1220.106494] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b0c6cd-e5c7-42bf-a826-d73cf266741c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.114439] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21df3b56-8a76-4dc3-97be-b3f9d8d9be01 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.211579] env[61947]: DEBUG nova.compute.manager [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Received event network-vif-plugged-9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1220.211967] env[61947]: DEBUG oslo_concurrency.lockutils [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] Acquiring lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.212216] env[61947]: DEBUG oslo_concurrency.lockutils [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.212394] env[61947]: DEBUG oslo_concurrency.lockutils [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.212569] env[61947]: DEBUG nova.compute.manager [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] No waiting events found dispatching network-vif-plugged-9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1220.212737] env[61947]: WARNING nova.compute.manager [req-c2e9614d-65b9-4f6d-bf7f-570fe1084210 req-948bbd46-8cd8-49a2-a045-04adf58f5fd9 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Received unexpected event network-vif-plugged-9c7403f4-d529-4b98-be43-36b21c5c5294 for instance with vm_state building and task_state spawning. [ 1220.292554] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Successfully updated port: 9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1220.795739] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.795880] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.796123] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1221.336799] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1221.463406] env[61947]: DEBUG nova.network.neutron [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Updating instance_info_cache with network_info: [{"id": "9c7403f4-d529-4b98-be43-36b21c5c5294", "address": "fa:16:3e:0a:f6:60", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c7403f4-d5", "ovs_interfaceid": "9c7403f4-d529-4b98-be43-36b21c5c5294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.150533] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.150819] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Instance network_info: |[{"id": "9c7403f4-d529-4b98-be43-36b21c5c5294", "address": "fa:16:3e:0a:f6:60", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c7403f4-d5", "ovs_interfaceid": "9c7403f4-d529-4b98-be43-36b21c5c5294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1222.151244] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:f6:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c7403f4-d529-4b98-be43-36b21c5c5294', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1222.158422] env[61947]: DEBUG oslo.service.loopingcall [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1222.158907] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1222.159148] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8cba098-df8b-41ad-8496-a0ab826af8d2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.179406] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1222.179406] env[61947]: value = "task-1225084" [ 1222.179406] env[61947]: _type = "Task" [ 1222.179406] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.187626] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225084, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.240307] env[61947]: DEBUG nova.compute.manager [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Received event network-changed-9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1222.240307] env[61947]: DEBUG nova.compute.manager [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Refreshing instance network info cache due to event network-changed-9c7403f4-d529-4b98-be43-36b21c5c5294. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1222.240307] env[61947]: DEBUG oslo_concurrency.lockutils [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] Acquiring lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.240438] env[61947]: DEBUG oslo_concurrency.lockutils [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] Acquired lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.240603] env[61947]: DEBUG nova.network.neutron [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Refreshing network info cache for port 9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1222.689151] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225084, 'name': CreateVM_Task, 'duration_secs': 0.290823} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.689526] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1222.689984] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.690175] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.690499] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1222.690748] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95b2ddd8-f961-4693-87a6-a3e391646507 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.695123] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1222.695123] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5212d722-b9f3-84b3-1131-5c4c4c839318" [ 1222.695123] env[61947]: _type = "Task" [ 1222.695123] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.702433] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5212d722-b9f3-84b3-1131-5c4c4c839318, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.113044] env[61947]: DEBUG nova.network.neutron [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Updated VIF entry in instance network info cache for port 9c7403f4-d529-4b98-be43-36b21c5c5294. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1223.113415] env[61947]: DEBUG nova.network.neutron [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Updating instance_info_cache with network_info: [{"id": "9c7403f4-d529-4b98-be43-36b21c5c5294", "address": "fa:16:3e:0a:f6:60", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c7403f4-d5", "ovs_interfaceid": "9c7403f4-d529-4b98-be43-36b21c5c5294", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.205637] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]5212d722-b9f3-84b3-1131-5c4c4c839318, 'name': SearchDatastore_Task, 'duration_secs': 0.009061} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.205886] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.206151] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1223.206396] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1223.206550] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.206731] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1223.206994] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6dbb1bd-c83f-4189-846e-c7ceb4146a55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.214586] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1223.214763] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1223.215468] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e744f3-5f9a-41bc-aa34-ef9c6cf13e84 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.220097] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1223.220097] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528f9bd7-12bd-e96d-a503-5aef42955ffd" [ 1223.220097] env[61947]: _type = "Task" [ 1223.220097] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.227627] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528f9bd7-12bd-e96d-a503-5aef42955ffd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.616224] env[61947]: DEBUG oslo_concurrency.lockutils [req-b13f49e7-1461-4839-94f1-d78b26e82751 req-f6c25456-884b-42f9-92f7-70ae92e6509c service nova] Releasing lock "refresh_cache-5bc960ad-3a21-465e-b6b8-83f04064745c" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.730825] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]528f9bd7-12bd-e96d-a503-5aef42955ffd, 'name': SearchDatastore_Task, 'duration_secs': 0.007894} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.731617] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c327656-eacc-4f10-a614-321da420a01d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.736543] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1223.736543] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524a9946-666f-5d29-528f-e16583db39c4" [ 1223.736543] env[61947]: _type = "Task" [ 1223.736543] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.744037] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524a9946-666f-5d29-528f-e16583db39c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.247380] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]524a9946-666f-5d29-528f-e16583db39c4, 'name': SearchDatastore_Task, 'duration_secs': 0.008898} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.247626] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.247918] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 5bc960ad-3a21-465e-b6b8-83f04064745c/5bc960ad-3a21-465e-b6b8-83f04064745c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1224.248194] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76bdd961-cb70-416e-9fbc-aa9b9ba46063 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.254080] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1224.254080] env[61947]: value = "task-1225085" [ 1224.254080] env[61947]: _type = "Task" [ 1224.254080] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.260898] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.763805] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225085, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442119} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.764192] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] 5bc960ad-3a21-465e-b6b8-83f04064745c/5bc960ad-3a21-465e-b6b8-83f04064745c.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1224.764320] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1224.764571] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02c55958-ea4e-412f-aab7-a11f92bd4640 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.771206] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1224.771206] env[61947]: value = "task-1225086" [ 1224.771206] env[61947]: _type = "Task" [ 1224.771206] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.777907] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.280790] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073763} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.281058] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1225.281807] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e60c539-249e-48d4-ba65-5566ac03c533 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.302295] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 5bc960ad-3a21-465e-b6b8-83f04064745c/5bc960ad-3a21-465e-b6b8-83f04064745c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1225.302526] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63f4fad9-8ee6-4cce-95c4-83a2fe99da45 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.320525] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1225.320525] env[61947]: value = "task-1225087" [ 1225.320525] env[61947]: _type = "Task" [ 1225.320525] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.327521] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225087, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.831459] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225087, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.332361] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225087, 'name': ReconfigVM_Task, 'duration_secs': 0.546762} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.332361] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 5bc960ad-3a21-465e-b6b8-83f04064745c/5bc960ad-3a21-465e-b6b8-83f04064745c.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1226.332592] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b5d7190-04ba-4268-9714-edb51f05a983 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.338601] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1226.338601] env[61947]: value = "task-1225088" [ 1226.338601] env[61947]: _type = "Task" [ 1226.338601] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.346099] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225088, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.848695] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225088, 'name': Rename_Task, 'duration_secs': 0.130198} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.851046] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1226.851046] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-519af7e7-a876-42c4-9dc5-2353925b12ae {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.855439] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1226.855439] env[61947]: value = "task-1225089" [ 1226.855439] env[61947]: _type = "Task" [ 1226.855439] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.862433] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.364899] env[61947]: DEBUG oslo_vmware.api [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225089, 'name': PowerOnVM_Task, 'duration_secs': 0.409145} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.365183] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1227.365424] env[61947]: INFO nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1227.365657] env[61947]: DEBUG nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1227.366455] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272f8683-1ae4-43e9-a00c-1f55e605ef14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.882713] env[61947]: INFO nova.compute.manager [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Took 11.95 seconds to build instance. [ 1228.385242] env[61947]: DEBUG oslo_concurrency.lockutils [None req-83e9ca15-475f-42db-b8a6-a391130df084 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.464s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.258155] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "5bc960ad-3a21-465e-b6b8-83f04064745c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.258493] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.258682] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.258872] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.259061] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.261535] env[61947]: INFO nova.compute.manager [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Terminating instance [ 1229.263265] env[61947]: DEBUG nova.compute.manager [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1229.263463] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1229.264343] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebbdf2e-9c2f-47e4-af17-1dc9fc1b9967 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.271620] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1229.272134] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cc4343a-28f1-4641-acbd-05bc194f2e6e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.278090] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1229.278090] env[61947]: value = "task-1225090" [ 1229.278090] env[61947]: _type = "Task" [ 1229.278090] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.285253] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.788531] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225090, 'name': PowerOffVM_Task, 'duration_secs': 0.190382} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.788795] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1229.788990] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1229.789270] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e671edf-1fa6-40b3-b61f-d4ba2a37a898 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.849278] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1229.849522] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1229.849691] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] 5bc960ad-3a21-465e-b6b8-83f04064745c {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1229.849950] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af029186-477d-4715-a965-96d640161a7d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.856149] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1229.856149] env[61947]: value = "task-1225092" [ 1229.856149] env[61947]: _type = "Task" [ 1229.856149] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.864926] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225092, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.365404] env[61947]: DEBUG oslo_vmware.api [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225092, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136275} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.365827] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.365827] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1230.366000] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1230.366203] env[61947]: INFO nova.compute.manager [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1230.366438] env[61947]: DEBUG oslo.service.loopingcall [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1230.366626] env[61947]: DEBUG nova.compute.manager [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1230.366712] env[61947]: DEBUG nova.network.neutron [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1230.592795] env[61947]: DEBUG nova.compute.manager [req-0a5e4972-993c-428d-a745-2a6efed37817 req-8d282c21-b0be-4690-8b16-34fbc92c5228 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Received event network-vif-deleted-9c7403f4-d529-4b98-be43-36b21c5c5294 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1230.592927] env[61947]: INFO nova.compute.manager [req-0a5e4972-993c-428d-a745-2a6efed37817 req-8d282c21-b0be-4690-8b16-34fbc92c5228 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Neutron deleted interface 9c7403f4-d529-4b98-be43-36b21c5c5294; detaching it from the instance and deleting it from the info cache [ 1230.593118] env[61947]: DEBUG nova.network.neutron [req-0a5e4972-993c-428d-a745-2a6efed37817 req-8d282c21-b0be-4690-8b16-34fbc92c5228 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.069322] env[61947]: DEBUG nova.network.neutron [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.095616] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bdd5756-d2bb-43b8-af53-6a2ba6effca4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.105523] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc667d80-e3ce-44d1-9f34-6037af72f8f9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.127735] env[61947]: DEBUG nova.compute.manager [req-0a5e4972-993c-428d-a745-2a6efed37817 req-8d282c21-b0be-4690-8b16-34fbc92c5228 service nova] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Detach interface failed, port_id=9c7403f4-d529-4b98-be43-36b21c5c5294, reason: Instance 5bc960ad-3a21-465e-b6b8-83f04064745c could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1231.571704] env[61947]: INFO nova.compute.manager [-] [instance: 5bc960ad-3a21-465e-b6b8-83f04064745c] Took 1.20 seconds to deallocate network for instance. [ 1232.077984] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.078338] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.078565] env[61947]: DEBUG nova.objects.instance [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid 5bc960ad-3a21-465e-b6b8-83f04064745c {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.620348] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be5f9ea-173a-4a37-b5c4-9961190b42b1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.627821] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aafc074-bf9c-4cf3-894e-d7f13752e917 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.656534] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9391d070-3937-442f-9910-aa93d58ae228 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.662866] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7f7669-e62c-4257-88d6-a5add8b98fd7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.675261] env[61947]: DEBUG nova.compute.provider_tree [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.178426] env[61947]: DEBUG nova.scheduler.client.report [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.683796] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.605s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.706460] env[61947]: INFO nova.scheduler.client.report [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance 5bc960ad-3a21-465e-b6b8-83f04064745c [ 1234.214886] env[61947]: DEBUG oslo_concurrency.lockutils [None req-2f3b7f5c-29f4-4a81-8028-bd8cc4ffe380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "5bc960ad-3a21-465e-b6b8-83f04064745c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.956s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.387390] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1234.387390] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1234.387390] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1234.916429] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.916700] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.916743] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1234.916879] env[61947]: DEBUG nova.objects.instance [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lazy-loading 'info_cache' on Instance uuid ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.165895] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "acc1df6b-150a-4305-be92-68d371b3ef04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.166110] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.669075] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1236.190282] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.190553] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.192036] env[61947]: INFO nova.compute.claims [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.628986] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [{"id": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "address": "fa:16:3e:b9:f2:a9", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16b0fd1c-1d", "ovs_interfaceid": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.132133] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.132401] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1237.132592] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1237.234726] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e249994-815d-4f13-b15e-430e3a560ec4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.242231] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79dfa79-3a4a-4a33-8e03-f77a70db38b0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.271305] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e21ab7-f386-44c8-ab89-cd24e0152bb9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.277856] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4249a0e9-bcd3-40d2-9422-70de3a4d3f6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.290017] env[61947]: DEBUG nova.compute.provider_tree [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.635514] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.792861] env[61947]: DEBUG nova.scheduler.client.report [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1238.298365] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.298889] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1238.301407] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.666s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.301592] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.301742] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1238.302790] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609b3a6f-6ed5-4772-9af9-ad1693cad061 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.311333] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992c8ae6-109f-493f-990e-d68f46329ba2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.324813] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cfa1d5-62b9-420f-8894-cf205a65a57c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.330876] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239f41f7-5fa6-4231-95e9-343d5cfea02f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.360263] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181490MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1238.360433] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.360599] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.804964] env[61947]: DEBUG nova.compute.utils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1238.806338] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1238.806515] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1238.855747] env[61947]: DEBUG nova.policy [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1239.087270] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Successfully created port: 6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1239.309290] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1239.382873] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1239.383049] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance acc1df6b-150a-4305-be92-68d371b3ef04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1239.383235] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1239.383373] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1239.420260] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ecf031-e325-4e09-9a44-382fd2404ada {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.427759] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d49c1f7-a932-4bbd-a8e9-d41e52b0c652 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.456330] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b587307-c52e-42b2-a030-070c7497c22b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.463194] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7989796d-e482-4f5c-b472-a7ab4826d82b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.476793] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.979647] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1240.319329] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1240.343530] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1240.343789] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1240.343951] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1240.344181] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1240.344339] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1240.344489] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1240.344693] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1240.344854] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1240.345060] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1240.345292] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1240.345481] env[61947]: DEBUG nova.virt.hardware [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.346372] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186cc0c1-3c50-4ce5-ac37-d3fd6683e14a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.354185] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276cc134-a6ff-4bef-986a-caeb0f95b997 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.454650] env[61947]: DEBUG nova.compute.manager [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Received event network-vif-plugged-6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1240.454880] env[61947]: DEBUG oslo_concurrency.lockutils [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] Acquiring lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.455137] env[61947]: DEBUG oslo_concurrency.lockutils [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] Lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.455321] env[61947]: DEBUG oslo_concurrency.lockutils [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] Lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.455490] env[61947]: DEBUG nova.compute.manager [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] No waiting events found dispatching network-vif-plugged-6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1240.455657] env[61947]: WARNING nova.compute.manager [req-ef82efeb-10fc-489f-87a1-45f1da92eeab req-5c666c4d-8319-4c98-b4d3-97d8c53c5503 service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Received unexpected event network-vif-plugged-6dec1590-2c8c-4244-9375-9fdce6e3b761 for instance with vm_state building and task_state spawning. [ 1240.484824] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1240.485011] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.534274] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Successfully updated port: 6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1241.036873] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.037138] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.037196] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1241.568097] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1241.683091] env[61947]: DEBUG nova.network.neutron [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Updating instance_info_cache with network_info: [{"id": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "address": "fa:16:3e:30:3c:86", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dec1590-2c", "ovs_interfaceid": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.738055] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738055] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738242] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738242] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738395] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738538] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.738674] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1242.186068] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.186415] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Instance network_info: |[{"id": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "address": "fa:16:3e:30:3c:86", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dec1590-2c", "ovs_interfaceid": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1242.187051] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:3c:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6dec1590-2c8c-4244-9375-9fdce6e3b761', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1242.194213] env[61947]: DEBUG oslo.service.loopingcall [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1242.194437] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1242.194655] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-692f5a83-48c7-44fd-84ed-19b37909e066 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.214240] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1242.214240] env[61947]: value = "task-1225093" [ 1242.214240] env[61947]: _type = "Task" [ 1242.214240] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.223470] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225093, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.386493] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.483777] env[61947]: DEBUG nova.compute.manager [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Received event network-changed-6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1242.483860] env[61947]: DEBUG nova.compute.manager [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Refreshing instance network info cache due to event network-changed-6dec1590-2c8c-4244-9375-9fdce6e3b761. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1242.484134] env[61947]: DEBUG oslo_concurrency.lockutils [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] Acquiring lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.484305] env[61947]: DEBUG oslo_concurrency.lockutils [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] Acquired lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.484485] env[61947]: DEBUG nova.network.neutron [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Refreshing network info cache for port 6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1242.725122] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225093, 'name': CreateVM_Task, 'duration_secs': 0.292089} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.725122] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1242.725619] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.725619] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.725794] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1242.726066] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e01dffc4-208d-4907-aaad-3df8962dea5c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.730874] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1242.730874] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52797fed-fdcb-3901-ed0c-cdb715f219f8" [ 1242.730874] env[61947]: _type = "Task" [ 1242.730874] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.738371] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52797fed-fdcb-3901-ed0c-cdb715f219f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.175901] env[61947]: DEBUG nova.network.neutron [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Updated VIF entry in instance network info cache for port 6dec1590-2c8c-4244-9375-9fdce6e3b761. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1243.176289] env[61947]: DEBUG nova.network.neutron [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Updating instance_info_cache with network_info: [{"id": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "address": "fa:16:3e:30:3c:86", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dec1590-2c", "ovs_interfaceid": "6dec1590-2c8c-4244-9375-9fdce6e3b761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.241647] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52797fed-fdcb-3901-ed0c-cdb715f219f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008778} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.241970] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.242245] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1243.242484] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.242638] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.242814] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1243.243089] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2c1bc64-2e05-4ae1-8577-4fd36f3ad469 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.250856] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1243.251063] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1243.251749] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06e442f0-b56e-405e-8f3b-31f0f31bde28 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.256389] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1243.256389] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521df189-a7c5-db1e-42b4-2631e7141a33" [ 1243.256389] env[61947]: _type = "Task" [ 1243.256389] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.263863] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521df189-a7c5-db1e-42b4-2631e7141a33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.679190] env[61947]: DEBUG oslo_concurrency.lockutils [req-012fbe7c-1498-4a53-8e44-83b5282b4c9c req-d8b2069e-b67f-4f92-a535-815f00d1468d service nova] Releasing lock "refresh_cache-acc1df6b-150a-4305-be92-68d371b3ef04" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.766581] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521df189-a7c5-db1e-42b4-2631e7141a33, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.767348] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8da2b30-e514-4195-a298-37b205e848f8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.771833] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1243.771833] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bb480-deb3-1d3e-02d9-1aa431a68727" [ 1243.771833] env[61947]: _type = "Task" [ 1243.771833] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.778988] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bb480-deb3-1d3e-02d9-1aa431a68727, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.282727] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]520bb480-deb3-1d3e-02d9-1aa431a68727, 'name': SearchDatastore_Task, 'duration_secs': 0.008402} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.283047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.283368] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] acc1df6b-150a-4305-be92-68d371b3ef04/acc1df6b-150a-4305-be92-68d371b3ef04.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1244.283661] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c240ea0c-4df8-40b5-80e6-0ecdf0a4cd9d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.289811] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1244.289811] env[61947]: value = "task-1225094" [ 1244.289811] env[61947]: _type = "Task" [ 1244.289811] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.297511] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.800177] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.406965} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.800528] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] acc1df6b-150a-4305-be92-68d371b3ef04/acc1df6b-150a-4305-be92-68d371b3ef04.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1244.800596] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1244.800848] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c62ad0cd-eb8f-49ad-9178-6c177732603d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.806766] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1244.806766] env[61947]: value = "task-1225095" [ 1244.806766] env[61947]: _type = "Task" [ 1244.806766] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.813504] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.316614] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059753} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.316890] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1245.317657] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2739b7f4-2d61-4ed1-8a77-7d4ceff995d0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.339328] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] acc1df6b-150a-4305-be92-68d371b3ef04/acc1df6b-150a-4305-be92-68d371b3ef04.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1245.339570] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fde40b2-e7ac-4f27-b491-40072b638b13 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.358389] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1245.358389] env[61947]: value = "task-1225096" [ 1245.358389] env[61947]: _type = "Task" [ 1245.358389] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.366012] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225096, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.868835] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225096, 'name': ReconfigVM_Task, 'duration_secs': 0.342243} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.869218] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Reconfigured VM instance instance-0000006a to attach disk [datastore2] acc1df6b-150a-4305-be92-68d371b3ef04/acc1df6b-150a-4305-be92-68d371b3ef04.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1245.869791] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9df4e21e-911b-46d9-90d8-16d48c8bd901 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.875542] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1245.875542] env[61947]: value = "task-1225097" [ 1245.875542] env[61947]: _type = "Task" [ 1245.875542] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.882760] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225097, 'name': Rename_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.385749] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225097, 'name': Rename_Task, 'duration_secs': 0.133621} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.386084] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1246.386359] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9cba819-8184-4c62-b1ed-7d0eeaf2c9a0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.392953] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1246.392953] env[61947]: value = "task-1225098" [ 1246.392953] env[61947]: _type = "Task" [ 1246.392953] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.400453] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.903474] env[61947]: DEBUG oslo_vmware.api [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225098, 'name': PowerOnVM_Task, 'duration_secs': 0.419804} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.903823] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1246.903913] env[61947]: INFO nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Took 6.58 seconds to spawn the instance on the hypervisor. [ 1246.904105] env[61947]: DEBUG nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1246.904865] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1b7b66-9b56-46b1-b765-cfcbb79652ac {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.381017] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.420261] env[61947]: INFO nova.compute.manager [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Took 11.25 seconds to build instance. [ 1247.922680] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c8ebf041-b1e5-4f53-b6f4-d67446022d2e tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.756s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.671975] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.672242] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.174419] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1249.692986] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.693278] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.694846] env[61947]: INFO nova.compute.claims [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1250.745644] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9988b5-f664-419b-a2df-12d0a74600f3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.753197] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399135d3-ebcd-4b04-8170-f992f717e5c3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.781315] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c5851b-aedf-4096-81e0-7c94deb69f44 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.787968] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d08fe7c-17a5-4e1d-a64b-0529c9f8a199 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.800241] env[61947]: DEBUG nova.compute.provider_tree [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.303462] env[61947]: DEBUG nova.scheduler.client.report [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1251.808720] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.115s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.809295] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1252.316232] env[61947]: DEBUG nova.compute.utils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1252.318057] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1252.318205] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1252.401895] env[61947]: DEBUG nova.policy [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1252.671139] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Successfully created port: e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1252.821497] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1253.831653] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1253.856983] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1253.857268] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1253.857439] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1253.857630] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1253.857781] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1253.857932] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1253.858164] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1253.858330] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1253.858505] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1253.858671] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1253.858845] env[61947]: DEBUG nova.virt.hardware [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1253.859729] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7914d3-b601-4e32-a7e8-74e09018a51d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.868078] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4409150-2c25-4f51-bee7-13d1bbdcde23 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.025022] env[61947]: DEBUG nova.compute.manager [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Received event network-vif-plugged-e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1254.025275] env[61947]: DEBUG oslo_concurrency.lockutils [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] Acquiring lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.025485] env[61947]: DEBUG oslo_concurrency.lockutils [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.025717] env[61947]: DEBUG oslo_concurrency.lockutils [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.026027] env[61947]: DEBUG nova.compute.manager [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] No waiting events found dispatching network-vif-plugged-e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1254.026352] env[61947]: WARNING nova.compute.manager [req-c6b03bd3-97a3-4059-a33b-73e4965effbe req-9cd69419-3d2b-4649-a6dd-35a36bfc662c service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Received unexpected event network-vif-plugged-e54bbab5-7660-4c91-885f-a14bbb7811fd for instance with vm_state building and task_state spawning. [ 1254.103747] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Successfully updated port: e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1254.606403] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.606403] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.606403] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1255.137433] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1255.256935] env[61947]: DEBUG nova.network.neutron [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Updating instance_info_cache with network_info: [{"id": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "address": "fa:16:3e:b1:75:0f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54bbab5-76", "ovs_interfaceid": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.759637] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.759989] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Instance network_info: |[{"id": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "address": "fa:16:3e:b1:75:0f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54bbab5-76", "ovs_interfaceid": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1255.760448] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:75:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e54bbab5-7660-4c91-885f-a14bbb7811fd', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1255.767999] env[61947]: DEBUG oslo.service.loopingcall [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1255.768235] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1255.768907] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-839bcc4c-51a8-4cf6-97ea-01d11dce44ee {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.788336] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1255.788336] env[61947]: value = "task-1225099" [ 1255.788336] env[61947]: _type = "Task" [ 1255.788336] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.796465] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225099, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.051534] env[61947]: DEBUG nova.compute.manager [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Received event network-changed-e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1256.051743] env[61947]: DEBUG nova.compute.manager [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Refreshing instance network info cache due to event network-changed-e54bbab5-7660-4c91-885f-a14bbb7811fd. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1256.051980] env[61947]: DEBUG oslo_concurrency.lockutils [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] Acquiring lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.052160] env[61947]: DEBUG oslo_concurrency.lockutils [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] Acquired lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.052343] env[61947]: DEBUG nova.network.neutron [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Refreshing network info cache for port e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1256.298009] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225099, 'name': CreateVM_Task, 'duration_secs': 0.270771} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.298344] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1256.298802] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.298982] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.299333] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1256.299587] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ce51b71-3ece-4987-8535-22668b0fadcf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.303778] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1256.303778] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52980af0-cb7e-2a92-174a-de3d30115605" [ 1256.303778] env[61947]: _type = "Task" [ 1256.303778] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.311815] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52980af0-cb7e-2a92-174a-de3d30115605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.732703] env[61947]: DEBUG nova.network.neutron [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Updated VIF entry in instance network info cache for port e54bbab5-7660-4c91-885f-a14bbb7811fd. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1256.733087] env[61947]: DEBUG nova.network.neutron [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Updating instance_info_cache with network_info: [{"id": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "address": "fa:16:3e:b1:75:0f", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape54bbab5-76", "ovs_interfaceid": "e54bbab5-7660-4c91-885f-a14bbb7811fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.813891] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52980af0-cb7e-2a92-174a-de3d30115605, 'name': SearchDatastore_Task, 'duration_secs': 0.010155} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.815047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.815047] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1256.815047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.815047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.815268] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1256.815491] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66228f65-7289-480c-af35-d74dc8b08d09 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.823146] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1256.823319] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1256.824012] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6644399f-1f3b-49e7-81d8-eadd2b78b567 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.828523] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1256.828523] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c8d802-48dd-fc68-011a-2ed7d025128b" [ 1256.828523] env[61947]: _type = "Task" [ 1256.828523] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.835235] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c8d802-48dd-fc68-011a-2ed7d025128b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.235755] env[61947]: DEBUG oslo_concurrency.lockutils [req-fcd9eaab-180f-4a54-9903-cff682f0a8e7 req-c8444551-0e45-47aa-adf7-c2da02710f47 service nova] Releasing lock "refresh_cache-ac8bd31a-4edc-443a-a093-9b93fd4a35fd" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.339963] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52c8d802-48dd-fc68-011a-2ed7d025128b, 'name': SearchDatastore_Task, 'duration_secs': 0.008007} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.340748] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1fa18d3-a628-4acd-adb9-5fdadcbbfc97 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.345855] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1257.345855] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527cb926-41fc-d8ee-6bca-ca34ee4efb53" [ 1257.345855] env[61947]: _type = "Task" [ 1257.345855] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.352926] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527cb926-41fc-d8ee-6bca-ca34ee4efb53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.856506] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]527cb926-41fc-d8ee-6bca-ca34ee4efb53, 'name': SearchDatastore_Task, 'duration_secs': 0.009003} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.856789] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.857068] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ac8bd31a-4edc-443a-a093-9b93fd4a35fd/ac8bd31a-4edc-443a-a093-9b93fd4a35fd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1257.857334] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6914440e-be10-48ae-a624-c32b3d33d291 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.863827] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1257.863827] env[61947]: value = "task-1225100" [ 1257.863827] env[61947]: _type = "Task" [ 1257.863827] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.871014] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.373228] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451001} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.373683] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] ac8bd31a-4edc-443a-a093-9b93fd4a35fd/ac8bd31a-4edc-443a-a093-9b93fd4a35fd.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1258.373683] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1258.373916] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a2188cd-7cd6-4c51-a4ed-4c6674cf9166 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.380469] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1258.380469] env[61947]: value = "task-1225101" [ 1258.380469] env[61947]: _type = "Task" [ 1258.380469] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.387603] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.889605] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056409} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.889874] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1258.890628] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1b7eae-2bc9-4354-82ea-c63e53012737 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.912441] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] ac8bd31a-4edc-443a-a093-9b93fd4a35fd/ac8bd31a-4edc-443a-a093-9b93fd4a35fd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1258.912684] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6961f1a-229a-49bd-9545-9bfec2fb8411 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.931621] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1258.931621] env[61947]: value = "task-1225102" [ 1258.931621] env[61947]: _type = "Task" [ 1258.931621] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.939050] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.441502] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225102, 'name': ReconfigVM_Task, 'duration_secs': 0.277781} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.441879] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Reconfigured VM instance instance-0000006b to attach disk [datastore2] ac8bd31a-4edc-443a-a093-9b93fd4a35fd/ac8bd31a-4edc-443a-a093-9b93fd4a35fd.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1259.442580] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a97d32cc-0fec-4ce9-b666-8f669f4ff1a8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.450411] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1259.450411] env[61947]: value = "task-1225103" [ 1259.450411] env[61947]: _type = "Task" [ 1259.450411] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.457672] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225103, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.962819] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225103, 'name': Rename_Task, 'duration_secs': 0.136655} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.963216] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1259.965157] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a72541e-bc94-40a1-a7ae-4dc956dee083 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.969561] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1259.969561] env[61947]: value = "task-1225104" [ 1259.969561] env[61947]: _type = "Task" [ 1259.969561] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.976783] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.479140] env[61947]: DEBUG oslo_vmware.api [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225104, 'name': PowerOnVM_Task, 'duration_secs': 0.435212} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.479514] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1260.479607] env[61947]: INFO nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1260.479788] env[61947]: DEBUG nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1260.480571] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa7f06f-70fb-4e0c-a99b-a2cbcf4ca090 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.997020] env[61947]: INFO nova.compute.manager [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Took 11.32 seconds to build instance. [ 1261.498934] env[61947]: DEBUG oslo_concurrency.lockutils [None req-aac082be-ff08-4eca-a613-47936c2ef1a8 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.826s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.022352] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.022627] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.022852] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.023057] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.023244] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.025456] env[61947]: INFO nova.compute.manager [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Terminating instance [ 1262.027228] env[61947]: DEBUG nova.compute.manager [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1262.027432] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1262.028253] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba5a836-0be8-4202-be96-0bf92d24ae40 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.035730] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1262.035956] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8abd2c82-e5c8-41a9-af4f-097d96ac2ed3 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.042599] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1262.042599] env[61947]: value = "task-1225105" [ 1262.042599] env[61947]: _type = "Task" [ 1262.042599] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.050160] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.553572] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225105, 'name': PowerOffVM_Task, 'duration_secs': 0.170551} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.553925] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1262.554030] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1262.554360] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6502505b-738b-475e-8405-f5e6b7080eed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.613771] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1262.614056] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1262.614291] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] ac8bd31a-4edc-443a-a093-9b93fd4a35fd {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1262.614640] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-430537d4-2507-4a15-86ff-b89a222c8e87 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.621400] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1262.621400] env[61947]: value = "task-1225107" [ 1262.621400] env[61947]: _type = "Task" [ 1262.621400] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.629953] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225107, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.131164] env[61947]: DEBUG oslo_vmware.api [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131601} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.131439] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.131627] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1263.131803] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1263.131978] env[61947]: INFO nova.compute.manager [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1263.132240] env[61947]: DEBUG oslo.service.loopingcall [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1263.132442] env[61947]: DEBUG nova.compute.manager [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1263.132539] env[61947]: DEBUG nova.network.neutron [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1263.372266] env[61947]: DEBUG nova.compute.manager [req-edf88cd7-2d8e-4793-a594-edfe515d1883 req-4f060b0a-64d4-49c1-9108-4c38efaa9f0a service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Received event network-vif-deleted-e54bbab5-7660-4c91-885f-a14bbb7811fd {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1263.372266] env[61947]: INFO nova.compute.manager [req-edf88cd7-2d8e-4793-a594-edfe515d1883 req-4f060b0a-64d4-49c1-9108-4c38efaa9f0a service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Neutron deleted interface e54bbab5-7660-4c91-885f-a14bbb7811fd; detaching it from the instance and deleting it from the info cache [ 1263.372368] env[61947]: DEBUG nova.network.neutron [req-edf88cd7-2d8e-4793-a594-edfe515d1883 req-4f060b0a-64d4-49c1-9108-4c38efaa9f0a service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.850046] env[61947]: DEBUG nova.network.neutron [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.874658] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7883495-449b-4661-b947-9a147c028d98 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.884636] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e602a3f6-d29b-4a16-81f2-deb48177d748 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.908157] env[61947]: DEBUG nova.compute.manager [req-edf88cd7-2d8e-4793-a594-edfe515d1883 req-4f060b0a-64d4-49c1-9108-4c38efaa9f0a service nova] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Detach interface failed, port_id=e54bbab5-7660-4c91-885f-a14bbb7811fd, reason: Instance ac8bd31a-4edc-443a-a093-9b93fd4a35fd could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1264.352947] env[61947]: INFO nova.compute.manager [-] [instance: ac8bd31a-4edc-443a-a093-9b93fd4a35fd] Took 1.22 seconds to deallocate network for instance. [ 1264.859908] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.860224] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.860449] env[61947]: DEBUG nova.objects.instance [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid ac8bd31a-4edc-443a-a093-9b93fd4a35fd {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.415399] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085d8013-a2a0-4487-95a1-17e2aa0536c4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.422767] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3365e6-9c80-4239-8532-e31c05e667d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.452935] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e50381-f50a-4f14-87b7-22dfacf98083 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.459757] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b4ca18-56c3-4550-bb97-c7776a8905ec {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.472393] env[61947]: DEBUG nova.compute.provider_tree [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1265.975952] env[61947]: DEBUG nova.scheduler.client.report [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1266.481434] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.498761] env[61947]: INFO nova.scheduler.client.report [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance ac8bd31a-4edc-443a-a093-9b93fd4a35fd [ 1267.005663] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e438dc00-ca9e-4e30-a6a8-c969aca61873 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ac8bd31a-4edc-443a-a093-9b93fd4a35fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.983s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.491365] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "acc1df6b-150a-4305-be92-68d371b3ef04" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.491687] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.491917] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.492125] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.492314] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.494471] env[61947]: INFO nova.compute.manager [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Terminating instance [ 1267.496184] env[61947]: DEBUG nova.compute.manager [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1267.496386] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1267.497239] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbbae3d-ddb6-4740-af20-970475c9faea {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.505159] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1267.505395] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6e649db-8292-45bc-8e2a-64728c61093a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.511105] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1267.511105] env[61947]: value = "task-1225108" [ 1267.511105] env[61947]: _type = "Task" [ 1267.511105] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.518948] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.022127] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225108, 'name': PowerOffVM_Task, 'duration_secs': 0.178836} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.022127] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1268.022127] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1268.022607] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0228dea-166e-4fd1-803e-244bf76dd950 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.085128] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1268.085354] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1268.085585] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] acc1df6b-150a-4305-be92-68d371b3ef04 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1268.085861] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b0fb1b9-4e2d-459c-a2ff-404ed540bf92 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.092606] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1268.092606] env[61947]: value = "task-1225110" [ 1268.092606] env[61947]: _type = "Task" [ 1268.092606] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.101408] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225110, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.602392] env[61947]: DEBUG oslo_vmware.api [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225110, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120794} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.602655] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1268.602845] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1268.603040] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1268.603234] env[61947]: INFO nova.compute.manager [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1268.603515] env[61947]: DEBUG oslo.service.loopingcall [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1268.603728] env[61947]: DEBUG nova.compute.manager [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1268.603824] env[61947]: DEBUG nova.network.neutron [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1268.824167] env[61947]: DEBUG nova.compute.manager [req-3261f679-c236-4425-a743-14b924b84f21 req-c7ec24bc-f5df-4420-a200-0e4c9b6d3bbd service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Received event network-vif-deleted-6dec1590-2c8c-4244-9375-9fdce6e3b761 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1268.824410] env[61947]: INFO nova.compute.manager [req-3261f679-c236-4425-a743-14b924b84f21 req-c7ec24bc-f5df-4420-a200-0e4c9b6d3bbd service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Neutron deleted interface 6dec1590-2c8c-4244-9375-9fdce6e3b761; detaching it from the instance and deleting it from the info cache [ 1268.824578] env[61947]: DEBUG nova.network.neutron [req-3261f679-c236-4425-a743-14b924b84f21 req-c7ec24bc-f5df-4420-a200-0e4c9b6d3bbd service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.304356] env[61947]: DEBUG nova.network.neutron [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.326897] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad6bbb9a-6d49-4571-8508-b8e9ea16fa24 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.338036] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66383d5-ae33-46e3-b4c8-528c4e3225fc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.360739] env[61947]: DEBUG nova.compute.manager [req-3261f679-c236-4425-a743-14b924b84f21 req-c7ec24bc-f5df-4420-a200-0e4c9b6d3bbd service nova] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Detach interface failed, port_id=6dec1590-2c8c-4244-9375-9fdce6e3b761, reason: Instance acc1df6b-150a-4305-be92-68d371b3ef04 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1269.807032] env[61947]: INFO nova.compute.manager [-] [instance: acc1df6b-150a-4305-be92-68d371b3ef04] Took 1.20 seconds to deallocate network for instance. [ 1270.313408] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.313734] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.313922] env[61947]: DEBUG nova.objects.instance [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid acc1df6b-150a-4305-be92-68d371b3ef04 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.855560] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172ceccf-4d46-47fb-ba39-b7ef0ad29872 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.863033] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56d879f-d99b-4924-8f47-26e1165d9cc9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.893302] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7e4bd1-dcc1-4085-82ce-c9aecfb70b0a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.899953] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3859668-7315-473b-9a55-4e69152075b6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.912912] env[61947]: DEBUG nova.compute.provider_tree [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.416762] env[61947]: DEBUG nova.scheduler.client.report [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.922157] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.939021] env[61947]: INFO nova.scheduler.client.report [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance acc1df6b-150a-4305-be92-68d371b3ef04 [ 1272.447137] env[61947]: DEBUG oslo_concurrency.lockutils [None req-e20b7a66-7553-4ba2-b3e7-9c9ec9f8df6a tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "acc1df6b-150a-4305-be92-68d371b3ef04" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.955s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.367831] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.368077] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.870175] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1274.390124] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1274.390399] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1274.391932] env[61947]: INFO nova.compute.claims [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1275.435116] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ed05d6-0b70-46dd-838c-f72815262949 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.443573] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62eb5e57-4b06-44d5-a925-59b3ca03616b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.471981] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156dcd6d-9d54-4938-a823-2586d852c9fb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.478624] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df81db83-dce4-4f14-bc43-0d79f2ff6d1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.490891] env[61947]: DEBUG nova.compute.provider_tree [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1275.993694] env[61947]: DEBUG nova.scheduler.client.report [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1276.498485] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.499138] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1277.003970] env[61947]: DEBUG nova.compute.utils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1277.005447] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1277.005621] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1277.052394] env[61947]: DEBUG nova.policy [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1277.294116] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Successfully created port: 01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1277.509697] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1278.519859] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1278.545269] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1278.545547] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1278.545711] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1278.545894] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1278.546066] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1278.546233] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1278.546441] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1278.546602] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1278.546771] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1278.546953] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1278.547173] env[61947]: DEBUG nova.virt.hardware [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1278.548052] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75d1c20-a4a0-4379-b1ad-742403d76bbe {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.556071] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c484502c-87b5-45d9-8281-c71387d9a194 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.654236] env[61947]: DEBUG nova.compute.manager [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Received event network-vif-plugged-01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1278.654470] env[61947]: DEBUG oslo_concurrency.lockutils [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] Acquiring lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.654703] env[61947]: DEBUG oslo_concurrency.lockutils [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.654886] env[61947]: DEBUG oslo_concurrency.lockutils [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.655192] env[61947]: DEBUG nova.compute.manager [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] No waiting events found dispatching network-vif-plugged-01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1278.655412] env[61947]: WARNING nova.compute.manager [req-8f16b977-7be8-4605-ba9a-c0e8e90d3983 req-6040bee4-9e05-4d9a-9f8b-08ff2098bab8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Received unexpected event network-vif-plugged-01618de8-0591-472d-9e63-6458cd265dcf for instance with vm_state building and task_state spawning. [ 1278.731979] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Successfully updated port: 01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1279.235119] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.235213] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.235426] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1279.770781] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1279.890050] env[61947]: DEBUG nova.network.neutron [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Updating instance_info_cache with network_info: [{"id": "01618de8-0591-472d-9e63-6458cd265dcf", "address": "fa:16:3e:69:1c:42", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01618de8-05", "ovs_interfaceid": "01618de8-0591-472d-9e63-6458cd265dcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1280.392854] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1280.393290] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Instance network_info: |[{"id": "01618de8-0591-472d-9e63-6458cd265dcf", "address": "fa:16:3e:69:1c:42", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01618de8-05", "ovs_interfaceid": "01618de8-0591-472d-9e63-6458cd265dcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1280.393752] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:1c:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01618de8-0591-472d-9e63-6458cd265dcf', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1280.401429] env[61947]: DEBUG oslo.service.loopingcall [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1280.401647] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1280.401871] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52c44afa-a299-45e0-9f7f-cf4335dca808 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.421102] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1280.421102] env[61947]: value = "task-1225111" [ 1280.421102] env[61947]: _type = "Task" [ 1280.421102] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.428577] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225111, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.682468] env[61947]: DEBUG nova.compute.manager [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Received event network-changed-01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1280.682723] env[61947]: DEBUG nova.compute.manager [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Refreshing instance network info cache due to event network-changed-01618de8-0591-472d-9e63-6458cd265dcf. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1280.682946] env[61947]: DEBUG oslo_concurrency.lockutils [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] Acquiring lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.683230] env[61947]: DEBUG oslo_concurrency.lockutils [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] Acquired lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.683423] env[61947]: DEBUG nova.network.neutron [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Refreshing network info cache for port 01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1280.931128] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225111, 'name': CreateVM_Task, 'duration_secs': 0.277448} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.931448] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1280.931959] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.932156] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.932486] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1280.932758] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0a6f435-7263-4bf4-870f-d32c9b6a9b99 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.936926] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1280.936926] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52577099-a258-fab5-b5d8-42dd4056c8ac" [ 1280.936926] env[61947]: _type = "Task" [ 1280.936926] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.944313] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52577099-a258-fab5-b5d8-42dd4056c8ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.361889] env[61947]: DEBUG nova.network.neutron [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Updated VIF entry in instance network info cache for port 01618de8-0591-472d-9e63-6458cd265dcf. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1281.362338] env[61947]: DEBUG nova.network.neutron [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Updating instance_info_cache with network_info: [{"id": "01618de8-0591-472d-9e63-6458cd265dcf", "address": "fa:16:3e:69:1c:42", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01618de8-05", "ovs_interfaceid": "01618de8-0591-472d-9e63-6458cd265dcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.447703] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52577099-a258-fab5-b5d8-42dd4056c8ac, 'name': SearchDatastore_Task, 'duration_secs': 0.008836} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.448047] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1281.448328] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1281.448615] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.448805] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.449049] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1281.449319] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ff2d5dc-f9f3-4d92-ade2-e387e3cd77e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.457440] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1281.457655] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1281.458530] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cf8b919-8e3c-4966-9746-84acec0940e1 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.463511] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1281.463511] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52105da4-d408-e7ea-f0bf-ee2f320b5e8e" [ 1281.463511] env[61947]: _type = "Task" [ 1281.463511] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.471215] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52105da4-d408-e7ea-f0bf-ee2f320b5e8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.865256] env[61947]: DEBUG oslo_concurrency.lockutils [req-d954c4df-cc3d-45f4-9e90-6e5c7012e7b0 req-ce8f4bd5-f20a-4e98-93cb-d00096a771d4 service nova] Releasing lock "refresh_cache-f9281ed9-22ee-45f7-86ed-5a570dd68f03" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1281.973775] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52105da4-d408-e7ea-f0bf-ee2f320b5e8e, 'name': SearchDatastore_Task, 'duration_secs': 0.007809} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.974543] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b8dcd70-ecdf-419a-a085-8cc79bc276f6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.979592] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1281.979592] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a4ce47-1383-08b0-fdef-74f036f4d387" [ 1281.979592] env[61947]: _type = "Task" [ 1281.979592] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.986829] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a4ce47-1383-08b0-fdef-74f036f4d387, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.489373] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52a4ce47-1383-08b0-fdef-74f036f4d387, 'name': SearchDatastore_Task, 'duration_secs': 0.008163} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.489708] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.490028] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] f9281ed9-22ee-45f7-86ed-5a570dd68f03/f9281ed9-22ee-45f7-86ed-5a570dd68f03.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1282.490334] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63d25b7c-108e-434f-b48c-531d6f06de1d {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.496669] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1282.496669] env[61947]: value = "task-1225112" [ 1282.496669] env[61947]: _type = "Task" [ 1282.496669] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.503645] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.006278] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412725} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.006623] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] f9281ed9-22ee-45f7-86ed-5a570dd68f03/f9281ed9-22ee-45f7-86ed-5a570dd68f03.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1283.006805] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1283.007073] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-992db18c-076d-41b3-86e1-8a9014bd2687 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.013386] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1283.013386] env[61947]: value = "task-1225113" [ 1283.013386] env[61947]: _type = "Task" [ 1283.013386] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.020915] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.522571] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064972} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.522887] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1283.523642] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48565c15-151f-4920-8801-d457f422a43e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.544194] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] f9281ed9-22ee-45f7-86ed-5a570dd68f03/f9281ed9-22ee-45f7-86ed-5a570dd68f03.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1283.544417] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6156ecc-978e-4dd6-b01e-ee05282d1d8f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.562433] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1283.562433] env[61947]: value = "task-1225114" [ 1283.562433] env[61947]: _type = "Task" [ 1283.562433] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.569200] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.072078] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225114, 'name': ReconfigVM_Task, 'duration_secs': 0.280459} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.072417] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Reconfigured VM instance instance-0000006c to attach disk [datastore2] f9281ed9-22ee-45f7-86ed-5a570dd68f03/f9281ed9-22ee-45f7-86ed-5a570dd68f03.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1284.073017] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13be1e80-728b-4775-8a2e-d684cdd6d317 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.078911] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1284.078911] env[61947]: value = "task-1225115" [ 1284.078911] env[61947]: _type = "Task" [ 1284.078911] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.087174] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225115, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.588590] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225115, 'name': Rename_Task, 'duration_secs': 0.151027} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.588860] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1284.589115] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ce8adac-bb27-4d89-bd5c-afe420ab5ef8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.594670] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1284.594670] env[61947]: value = "task-1225116" [ 1284.594670] env[61947]: _type = "Task" [ 1284.594670] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.601965] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.105292] env[61947]: DEBUG oslo_vmware.api [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225116, 'name': PowerOnVM_Task, 'duration_secs': 0.456285} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.105669] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1285.105797] env[61947]: INFO nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Took 6.59 seconds to spawn the instance on the hypervisor. [ 1285.105975] env[61947]: DEBUG nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1285.106741] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a89f17-e1cd-45f8-993b-5c4e07900570 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.624887] env[61947]: INFO nova.compute.manager [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Took 11.25 seconds to build instance. [ 1286.127350] env[61947]: DEBUG oslo_concurrency.lockutils [None req-5f3467de-0f12-480d-b77f-53e0b7f87827 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.759s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.766064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.766064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.766064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.766064] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.766372] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.768392] env[61947]: INFO nova.compute.manager [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Terminating instance [ 1286.770124] env[61947]: DEBUG nova.compute.manager [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1286.770329] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1286.771154] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75533355-cf19-42af-a839-2b6b5abc8350 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.778601] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1286.779099] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66f559a3-92e4-4691-a2ca-8c0f926dd4f2 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.784762] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1286.784762] env[61947]: value = "task-1225117" [ 1286.784762] env[61947]: _type = "Task" [ 1286.784762] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.792103] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.295165] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225117, 'name': PowerOffVM_Task, 'duration_secs': 0.17935} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.295524] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1287.295607] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1287.295883] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-269edbaa-ecc5-44c1-9ff1-38dfb6aea650 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.353926] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1287.354173] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1287.354360] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] f9281ed9-22ee-45f7-86ed-5a570dd68f03 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1287.354625] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff559976-48d6-4fee-ba7f-06fc8535430e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.360819] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1287.360819] env[61947]: value = "task-1225119" [ 1287.360819] env[61947]: _type = "Task" [ 1287.360819] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.368641] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225119, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.871618] env[61947]: DEBUG oslo_vmware.api [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225119, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118955} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.871881] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1287.872088] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1287.872273] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1287.872454] env[61947]: INFO nova.compute.manager [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1287.872706] env[61947]: DEBUG oslo.service.loopingcall [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1287.872924] env[61947]: DEBUG nova.compute.manager [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1287.873030] env[61947]: DEBUG nova.network.neutron [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1288.089565] env[61947]: DEBUG nova.compute.manager [req-acfb3e9c-320c-4f60-b12a-653674625f68 req-db8af074-1fb5-4534-bcae-5352b3c748e8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Received event network-vif-deleted-01618de8-0591-472d-9e63-6458cd265dcf {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1288.089791] env[61947]: INFO nova.compute.manager [req-acfb3e9c-320c-4f60-b12a-653674625f68 req-db8af074-1fb5-4534-bcae-5352b3c748e8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Neutron deleted interface 01618de8-0591-472d-9e63-6458cd265dcf; detaching it from the instance and deleting it from the info cache [ 1288.089947] env[61947]: DEBUG nova.network.neutron [req-acfb3e9c-320c-4f60-b12a-653674625f68 req-db8af074-1fb5-4534-bcae-5352b3c748e8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.568665] env[61947]: DEBUG nova.network.neutron [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.593152] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42c4568b-13d1-405b-97ee-9851fea5aeb4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.601749] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4289fa3d-fe08-46ba-aeb8-1d7364ec16cf {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.624434] env[61947]: DEBUG nova.compute.manager [req-acfb3e9c-320c-4f60-b12a-653674625f68 req-db8af074-1fb5-4534-bcae-5352b3c748e8 service nova] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Detach interface failed, port_id=01618de8-0591-472d-9e63-6458cd265dcf, reason: Instance f9281ed9-22ee-45f7-86ed-5a570dd68f03 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1289.071177] env[61947]: INFO nova.compute.manager [-] [instance: f9281ed9-22ee-45f7-86ed-5a570dd68f03] Took 1.20 seconds to deallocate network for instance. [ 1289.577992] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1289.578399] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1289.578706] env[61947]: DEBUG nova.objects.instance [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid f9281ed9-22ee-45f7-86ed-5a570dd68f03 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1290.122243] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3f3d73-f5db-4680-8a13-21b14a091b0c {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.129829] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af60558-0f79-450b-bc11-6f267bfb6632 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.158471] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4988bf-5ba3-4da0-8916-7a67ded326e4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.165500] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568d5fdd-cd26-4c03-b44e-ec84e956c542 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.179634] env[61947]: DEBUG nova.compute.provider_tree [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1290.682511] env[61947]: DEBUG nova.scheduler.client.report [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1291.187914] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.204975] env[61947]: INFO nova.scheduler.client.report [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance f9281ed9-22ee-45f7-86ed-5a570dd68f03 [ 1291.713807] env[61947]: DEBUG oslo_concurrency.lockutils [None req-04acadd5-0f5e-47c1-aaaa-1c63b8ee6380 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "f9281ed9-22ee-45f7-86ed-5a570dd68f03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.948s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.588572] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "dad9de8f-4958-444f-a677-0f8b11e01bed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.588864] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.090846] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Starting instance... {{(pid=61947) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1293.612866] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.613165] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.614749] env[61947]: INFO nova.compute.claims [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1294.385327] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1294.385746] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Starting heal instance info cache {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1294.385746] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Rebuilding the list of instances to heal {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1294.660462] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eafbb2a-172b-493b-8f21-054798dddff8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.669183] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2d8a1-156e-4f47-bc10-1130dde0127e {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.698151] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38764b7f-d7eb-4f3c-818a-3128eeed88d5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.705505] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c81bf2-8105-465a-a8a1-ff60eff2ff5a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.718710] env[61947]: DEBUG nova.compute.provider_tree [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1294.888681] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Skipping network cache update for instance because it is Building. {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1294.927210] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1294.927376] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquired lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1294.927524] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Forcefully refreshing network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1294.927683] env[61947]: DEBUG nova.objects.instance [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lazy-loading 'info_cache' on Instance uuid ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1295.221924] env[61947]: DEBUG nova.scheduler.client.report [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1295.726480] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.727089] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Start building networks asynchronously for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1296.231870] env[61947]: DEBUG nova.compute.utils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Using /dev/sd instead of None {{(pid=61947) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1296.235720] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Allocating IP information in the background. {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1296.235887] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] allocate_for_instance() {{(pid=61947) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1296.282653] env[61947]: DEBUG nova.policy [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c38857a7b9e4095a4c1bf3339bb9e97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ef945423f04aa6b5e50d13138d08a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61947) authorize /opt/stack/nova/nova/policy.py:203}} [ 1296.586308] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Successfully created port: 6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1296.736485] env[61947]: DEBUG nova.network.neutron [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [{"id": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "address": "fa:16:3e:b9:f2:a9", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16b0fd1c-1d", "ovs_interfaceid": "16b0fd1c-1dae-455c-ba72-c515ff774b9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1296.738440] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Start building block device mappings for instance. {{(pid=61947) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1297.239158] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Releasing lock "refresh_cache-ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1297.239383] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updated the network info_cache for instance {{(pid=61947) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1297.748052] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Start spawning the instance on the hypervisor. {{(pid=61947) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1297.772534] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T14:43:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T14:43:08Z,direct_url=,disk_format='vmdk',id=24a0d126-58d4-4780-b581-953c5b860675,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='7706c941d6d844f588f72bcf7d3ea118',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T14:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1297.772801] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1297.772998] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image limits 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1297.773211] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Flavor pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1297.773367] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Image pref 0:0:0 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1297.773517] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61947) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1297.773724] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1297.773886] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1297.774072] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Got 1 possible topologies {{(pid=61947) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1297.774241] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1297.774414] env[61947]: DEBUG nova.virt.hardware [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61947) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1297.775353] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a425393f-ad02-472f-a1ca-8472b52f05dd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.783241] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f2fe59-bc0c-4d73-99e1-096957ad5073 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.017067] env[61947]: DEBUG nova.compute.manager [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Received event network-vif-plugged-6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1298.017311] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] Acquiring lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.017530] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.017705] env[61947]: DEBUG oslo_concurrency.lockutils [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.017880] env[61947]: DEBUG nova.compute.manager [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] No waiting events found dispatching network-vif-plugged-6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1298.018104] env[61947]: WARNING nova.compute.manager [req-2edb3f9e-33b0-42b5-9991-ea208fbc30bb req-001a7522-72f0-4843-be53-5b0bf9665a9a service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Received unexpected event network-vif-plugged-6534bc7b-99f4-45cf-a86d-2c1b32ac1041 for instance with vm_state building and task_state spawning. [ 1298.099266] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Successfully updated port: 6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1298.384820] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1298.601921] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1298.602080] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1298.602182] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Building network info cache for instance {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1298.887717] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.888034] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.888147] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.888302] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61947) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1298.889552] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7356f534-fde9-4d83-9fa8-1dfe01b9c478 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.897805] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22a6b6b-3955-4ab6-8352-450483569646 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.911825] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d7538b-96d1-4aa6-a845-7e04c35326b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.917781] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0541b164-ce72-457f-8d81-fa2b6394846f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.946286] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181290MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=61947) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1298.946445] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.946617] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.132993] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Instance cache missing network info. {{(pid=61947) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1299.249702] env[61947]: DEBUG nova.network.neutron [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Updating instance_info_cache with network_info: [{"id": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "address": "fa:16:3e:1d:be:29", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6534bc7b-99", "ovs_interfaceid": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1299.752706] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1299.752997] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Instance network_info: |[{"id": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "address": "fa:16:3e:1d:be:29", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6534bc7b-99", "ovs_interfaceid": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61947) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1299.753458] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:be:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6534bc7b-99f4-45cf-a86d-2c1b32ac1041', 'vif_model': 'vmxnet3'}] {{(pid=61947) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1299.761215] env[61947]: DEBUG oslo.service.loopingcall [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1299.761426] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Creating VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1299.761648] env[61947]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e5ee2de-3121-497c-8a32-79883248d223 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.781870] env[61947]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1299.781870] env[61947]: value = "task-1225120" [ 1299.781870] env[61947]: _type = "Task" [ 1299.781870] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.789128] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225120, 'name': CreateVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.971476] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1299.971798] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Instance dad9de8f-4958-444f-a677-0f8b11e01bed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61947) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1299.971845] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1299.971979] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61947) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1300.009825] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8fdf42-05c0-465e-85f3-25fcf4b3d6b8 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.018808] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a43cf46-0972-4561-8cc8-c6a1f73e6157 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.049662] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5851e553-1065-4f29-b7d2-9d66f706bf99 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.053046] env[61947]: DEBUG nova.compute.manager [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Received event network-changed-6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1300.053244] env[61947]: DEBUG nova.compute.manager [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Refreshing instance network info cache due to event network-changed-6534bc7b-99f4-45cf-a86d-2c1b32ac1041. {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1300.053462] env[61947]: DEBUG oslo_concurrency.lockutils [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] Acquiring lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1300.053609] env[61947]: DEBUG oslo_concurrency.lockutils [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] Acquired lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1300.053770] env[61947]: DEBUG nova.network.neutron [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Refreshing network info cache for port 6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1300.060169] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e452db-c415-45ed-8560-2929a54621bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.073818] env[61947]: DEBUG nova.compute.provider_tree [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1300.291090] env[61947]: DEBUG oslo_vmware.api [-] Task: {'id': task-1225120, 'name': CreateVM_Task, 'duration_secs': 0.264215} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.291310] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Created VM on the ESX host {{(pid=61947) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1300.291906] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1300.292103] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1300.292443] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1300.292698] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bff3dfbf-9891-42af-9fe9-d15c2e81ba57 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.296688] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1300.296688] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ccabc1-e00a-06c7-b45c-980154d25b0e" [ 1300.296688] env[61947]: _type = "Task" [ 1300.296688] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.303446] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ccabc1-e00a-06c7-b45c-980154d25b0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.576849] env[61947]: DEBUG nova.scheduler.client.report [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1300.807037] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ccabc1-e00a-06c7-b45c-980154d25b0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008499} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.807371] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.807614] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Processing image 24a0d126-58d4-4780-b581-953c5b860675 {{(pid=61947) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1300.807851] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1300.808012] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1300.808202] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1300.808463] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4690beb5-0dc6-4baf-99ec-bb4617d684b5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.815897] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61947) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1300.816072] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61947) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1300.816746] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57af65ec-48c3-4875-957f-231b8808ff94 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.821364] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1300.821364] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521e214c-3825-8f10-241e-828af57530d4" [ 1300.821364] env[61947]: _type = "Task" [ 1300.821364] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.829074] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521e214c-3825-8f10-241e-828af57530d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.914769] env[61947]: DEBUG nova.network.neutron [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Updated VIF entry in instance network info cache for port 6534bc7b-99f4-45cf-a86d-2c1b32ac1041. {{(pid=61947) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1300.915279] env[61947]: DEBUG nova.network.neutron [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Updating instance_info_cache with network_info: [{"id": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "address": "fa:16:3e:1d:be:29", "network": {"id": "c9d8b6e9-fe0e-425d-af7b-f30a35b34240", "bridge": "br-int", "label": "tempest-ServersTestJSON-490393182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ef945423f04aa6b5e50d13138d08a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6534bc7b-99", "ovs_interfaceid": "6534bc7b-99f4-45cf-a86d-2c1b32ac1041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1301.080888] env[61947]: DEBUG nova.compute.resource_tracker [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61947) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1301.081291] env[61947]: DEBUG oslo_concurrency.lockutils [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.134s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.332223] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]521e214c-3825-8f10-241e-828af57530d4, 'name': SearchDatastore_Task, 'duration_secs': 0.007089} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.332932] env[61947]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7119917c-577f-4485-9055-a827f09f0826 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.337649] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1301.337649] env[61947]: value = "session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ecf8e5-3e33-ba01-9941-a62ac7b39b27" [ 1301.337649] env[61947]: _type = "Task" [ 1301.337649] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.344596] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ecf8e5-3e33-ba01-9941-a62ac7b39b27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.418297] env[61947]: DEBUG oslo_concurrency.lockutils [req-66af82ff-97dc-4b80-b240-fcadbd62325e req-19d08551-1c6d-4ccb-81ab-c79ebc8770df service nova] Releasing lock "refresh_cache-dad9de8f-4958-444f-a677-0f8b11e01bed" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1301.848513] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': session[52bd20a5-9519-69b8-30d5-dd490ecc5fa1]52ecf8e5-3e33-ba01-9941-a62ac7b39b27, 'name': SearchDatastore_Task, 'duration_secs': 0.008267} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.848846] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk" {{(pid=61947) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1301.849133] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] dad9de8f-4958-444f-a677-0f8b11e01bed/dad9de8f-4958-444f-a677-0f8b11e01bed.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1301.849392] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8caf969c-c0a0-413e-9af7-e7c4e805af25 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.855378] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1301.855378] env[61947]: value = "task-1225121" [ 1301.855378] env[61947]: _type = "Task" [ 1301.855378] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.862443] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.081639] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.082039] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.082039] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.082207] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.365013] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424321} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.365315] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24a0d126-58d4-4780-b581-953c5b860675/24a0d126-58d4-4780-b581-953c5b860675.vmdk to [datastore2] dad9de8f-4958-444f-a677-0f8b11e01bed/dad9de8f-4958-444f-a677-0f8b11e01bed.vmdk {{(pid=61947) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1302.365517] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Extending root virtual disk to 1048576 {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1302.365761] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1cfc17f-8a39-4cf2-b4a0-bf88a8afec55 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.371762] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1302.371762] env[61947]: value = "task-1225122" [ 1302.371762] env[61947]: _type = "Task" [ 1302.371762] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.379334] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.384858] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.385017] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.385196] env[61947]: DEBUG oslo_service.periodic_task [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61947) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.385322] env[61947]: DEBUG nova.compute.manager [None req-c7a35e7a-894f-419d-a6e7-ff7f83a07d46 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61947) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1302.881782] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070358} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.882088] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Extended root virtual disk {{(pid=61947) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1302.882855] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1f2859-641e-4af5-93ab-86c3f1244fb4 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.903914] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] dad9de8f-4958-444f-a677-0f8b11e01bed/dad9de8f-4958-444f-a677-0f8b11e01bed.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1302.904194] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94738e01-5c33-4b21-81ec-91345018a570 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.922988] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1302.922988] env[61947]: value = "task-1225123" [ 1302.922988] env[61947]: _type = "Task" [ 1302.922988] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.930227] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.432901] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225123, 'name': ReconfigVM_Task, 'duration_secs': 0.251838} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.433281] env[61947]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Reconfigured VM instance instance-0000006d to attach disk [datastore2] dad9de8f-4958-444f-a677-0f8b11e01bed/dad9de8f-4958-444f-a677-0f8b11e01bed.vmdk or device None with type sparse {{(pid=61947) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1303.433745] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ea2762c-afdd-48d6-a07f-555b0d6e8004 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.440326] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1303.440326] env[61947]: value = "task-1225124" [ 1303.440326] env[61947]: _type = "Task" [ 1303.440326] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.447864] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225124, 'name': Rename_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.950265] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225124, 'name': Rename_Task, 'duration_secs': 0.129022} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.950553] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Powering on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1303.950810] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4296a81b-128f-47f9-96f7-318613de5e14 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.957090] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1303.957090] env[61947]: value = "task-1225125" [ 1303.957090] env[61947]: _type = "Task" [ 1303.957090] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.965759] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.466679] env[61947]: DEBUG oslo_vmware.api [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225125, 'name': PowerOnVM_Task, 'duration_secs': 0.436749} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.467052] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Powered on the VM {{(pid=61947) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1304.467232] env[61947]: INFO nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Took 6.72 seconds to spawn the instance on the hypervisor. [ 1304.467364] env[61947]: DEBUG nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1304.468096] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea03559-eef5-458f-a96a-734f90af58cb {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.984854] env[61947]: INFO nova.compute.manager [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Took 11.39 seconds to build instance. [ 1305.487384] env[61947]: DEBUG oslo_concurrency.lockutils [None req-ae1fc54e-843f-4b15-889d-fcbbd5430fc2 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.898s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.301634] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "dad9de8f-4958-444f-a677-0f8b11e01bed" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.301853] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.302049] env[61947]: DEBUG nova.compute.manager [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1306.303438] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9475fa0b-dfad-4e38-8d47-43080293ea86 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.310313] env[61947]: DEBUG nova.compute.manager [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61947) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1306.310852] env[61947]: DEBUG nova.objects.instance [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'flavor' on Instance uuid dad9de8f-4958-444f-a677-0f8b11e01bed {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1306.816037] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1306.816398] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66d71d79-f7e2-4dc3-8dcf-39eb58b652bc {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.823747] env[61947]: DEBUG oslo_vmware.api [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1306.823747] env[61947]: value = "task-1225126" [ 1306.823747] env[61947]: _type = "Task" [ 1306.823747] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.831458] env[61947]: DEBUG oslo_vmware.api [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.332759] env[61947]: DEBUG oslo_vmware.api [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225126, 'name': PowerOffVM_Task, 'duration_secs': 0.188701} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.333064] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1307.333257] env[61947]: DEBUG nova.compute.manager [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Checking state {{(pid=61947) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1307.334018] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3817ed53-3c29-446b-b6ce-86fe82e09cd5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.845449] env[61947]: DEBUG oslo_concurrency.lockutils [None req-d07cd74f-1b1e-499c-8b4b-9f49e5b50337 tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1308.905728] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "dad9de8f-4958-444f-a677-0f8b11e01bed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.906171] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.906309] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.906508] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.906687] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1308.909087] env[61947]: INFO nova.compute.manager [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Terminating instance [ 1308.910826] env[61947]: DEBUG nova.compute.manager [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1308.911034] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1308.911873] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64429838-a4d2-445e-b6cf-791acd316060 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.919193] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1308.919430] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e91d421-a6d7-4a95-bb34-880dfe00aa27 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.973817] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1308.974046] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Deleting contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1308.974244] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore2] dad9de8f-4958-444f-a677-0f8b11e01bed {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1308.974504] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23c16e56-206d-4061-bf0e-2b161a63ebd0 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.980706] env[61947]: DEBUG oslo_vmware.api [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1308.980706] env[61947]: value = "task-1225128" [ 1308.980706] env[61947]: _type = "Task" [ 1308.980706] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.987934] env[61947]: DEBUG oslo_vmware.api [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.490511] env[61947]: DEBUG oslo_vmware.api [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12047} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.490774] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1309.490966] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Deleted contents of the VM from datastore datastore2 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1309.491168] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1309.491349] env[61947]: INFO nova.compute.manager [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Took 0.58 seconds to destroy the instance on the hypervisor. [ 1309.491592] env[61947]: DEBUG oslo.service.loopingcall [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1309.491785] env[61947]: DEBUG nova.compute.manager [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1309.491882] env[61947]: DEBUG nova.network.neutron [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1309.819720] env[61947]: DEBUG nova.compute.manager [req-073f211c-9d55-45a0-8c54-7fc5147dee1f req-35ad4479-7267-44b0-92b1-71f15c7b2c8f service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Received event network-vif-deleted-6534bc7b-99f4-45cf-a86d-2c1b32ac1041 {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1309.819919] env[61947]: INFO nova.compute.manager [req-073f211c-9d55-45a0-8c54-7fc5147dee1f req-35ad4479-7267-44b0-92b1-71f15c7b2c8f service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Neutron deleted interface 6534bc7b-99f4-45cf-a86d-2c1b32ac1041; detaching it from the instance and deleting it from the info cache [ 1309.820114] env[61947]: DEBUG nova.network.neutron [req-073f211c-9d55-45a0-8c54-7fc5147dee1f req-35ad4479-7267-44b0-92b1-71f15c7b2c8f service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.297466] env[61947]: DEBUG nova.network.neutron [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.323250] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe066913-e414-498f-97ae-35f9d98da804 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.331657] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177b2fd7-0a35-4cb8-adb3-b41590b358e6 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.353883] env[61947]: DEBUG nova.compute.manager [req-073f211c-9d55-45a0-8c54-7fc5147dee1f req-35ad4479-7267-44b0-92b1-71f15c7b2c8f service nova] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Detach interface failed, port_id=6534bc7b-99f4-45cf-a86d-2c1b32ac1041, reason: Instance dad9de8f-4958-444f-a677-0f8b11e01bed could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1310.800551] env[61947]: INFO nova.compute.manager [-] [instance: dad9de8f-4958-444f-a677-0f8b11e01bed] Took 1.31 seconds to deallocate network for instance. [ 1311.307063] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1311.307411] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.307632] env[61947]: DEBUG nova.objects.instance [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid dad9de8f-4958-444f-a677-0f8b11e01bed {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1311.851067] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb4ec86-11f8-4d43-999c-f8b038f64f6a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.860170] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3182df-1b8b-4c71-a4e6-d3d958610c02 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.890088] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0666d462-9a65-4aac-8eaa-5b45f8e715b7 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.897204] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dc6304-c0e4-42d2-9302-76b997d6a7e5 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.909922] env[61947]: DEBUG nova.compute.provider_tree [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1312.412754] env[61947]: DEBUG nova.scheduler.client.report [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1312.917510] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.935057] env[61947]: INFO nova.scheduler.client.report [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance dad9de8f-4958-444f-a677-0f8b11e01bed [ 1313.443660] env[61947]: DEBUG oslo_concurrency.lockutils [None req-72bdb2da-db20-4ef5-be93-00bcce2ccc7c tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "dad9de8f-4958-444f-a677-0f8b11e01bed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.538s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.679052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.679052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.679052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.679052] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.679450] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.681181] env[61947]: INFO nova.compute.manager [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Terminating instance [ 1314.682864] env[61947]: DEBUG nova.compute.manager [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Start destroying the instance on the hypervisor. {{(pid=61947) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1314.683099] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Destroying instance {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1314.683937] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb063a78-8d77-40cc-ab4d-b6abbaf9a5ad {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.691658] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Powering off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1314.691884] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4c75e9b-9336-4af1-9605-72771717fa1b {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.698281] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1314.698281] env[61947]: value = "task-1225129" [ 1314.698281] env[61947]: _type = "Task" [ 1314.698281] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.705820] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225129, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.208524] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225129, 'name': PowerOffVM_Task, 'duration_secs': 0.185218} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.208802] env[61947]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Powered off the VM {{(pid=61947) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1315.208974] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Unregistering the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1315.209235] env[61947]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51e47e49-e891-4fa5-a99c-c0b63eceeb17 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.269581] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Unregistered the VM {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1315.269818] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Deleting contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1315.269988] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleting the datastore file [datastore1] ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1315.270278] env[61947]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdde8cac-7e71-47c3-9526-772b567d73da {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.276567] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for the task: (returnval){ [ 1315.276567] env[61947]: value = "task-1225131" [ 1315.276567] env[61947]: _type = "Task" [ 1315.276567] env[61947]: } to complete. {{(pid=61947) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.283801] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225131, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.786695] env[61947]: DEBUG oslo_vmware.api [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Task: {'id': task-1225131, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121484} completed successfully. {{(pid=61947) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.787099] env[61947]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted the datastore file {{(pid=61947) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1315.787178] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Deleted contents of the VM from datastore datastore1 {{(pid=61947) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1315.787321] env[61947]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Instance destroyed {{(pid=61947) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1315.787552] env[61947]: INFO nova.compute.manager [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1315.787815] env[61947]: DEBUG oslo.service.loopingcall [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61947) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.788022] env[61947]: DEBUG nova.compute.manager [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Deallocating network for instance {{(pid=61947) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1315.788122] env[61947]: DEBUG nova.network.neutron [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] deallocate_for_instance() {{(pid=61947) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1316.014543] env[61947]: DEBUG nova.compute.manager [req-fb1fa834-a3fc-4607-b08f-31e943bbff70 req-7c1c3477-ddf1-49a0-b76f-96290610988b service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Received event network-vif-deleted-16b0fd1c-1dae-455c-ba72-c515ff774b9a {{(pid=61947) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1316.014727] env[61947]: INFO nova.compute.manager [req-fb1fa834-a3fc-4607-b08f-31e943bbff70 req-7c1c3477-ddf1-49a0-b76f-96290610988b service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Neutron deleted interface 16b0fd1c-1dae-455c-ba72-c515ff774b9a; detaching it from the instance and deleting it from the info cache [ 1316.014910] env[61947]: DEBUG nova.network.neutron [req-fb1fa834-a3fc-4607-b08f-31e943bbff70 req-7c1c3477-ddf1-49a0-b76f-96290610988b service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.494949] env[61947]: DEBUG nova.network.neutron [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Updating instance_info_cache with network_info: [] {{(pid=61947) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.517180] env[61947]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dcbc7e2-c27a-4c7b-b7a8-b7498216b6ed {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.526341] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95380191-4daa-4237-86ab-56616201c3cd {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.546890] env[61947]: DEBUG nova.compute.manager [req-fb1fa834-a3fc-4607-b08f-31e943bbff70 req-7c1c3477-ddf1-49a0-b76f-96290610988b service nova] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Detach interface failed, port_id=16b0fd1c-1dae-455c-ba72-c515ff774b9a, reason: Instance ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 could not be found. {{(pid=61947) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1316.998263] env[61947]: INFO nova.compute.manager [-] [instance: ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8] Took 1.21 seconds to deallocate network for instance. [ 1317.505328] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.505616] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.505842] env[61947]: DEBUG nova.objects.instance [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lazy-loading 'resources' on Instance uuid ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 {{(pid=61947) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1318.039580] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c80aee-a1fc-48db-856e-e49c6fe5fd9a {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.048613] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a02ca5-8c92-4fd3-8120-132b74071bb9 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.077290] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e25d72-2a7e-4ec2-9ef2-e1cef5b6e188 {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.083612] env[61947]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f3b684-5c48-486e-85ae-efc08e947b7f {{(pid=61947) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.095945] env[61947]: DEBUG nova.compute.provider_tree [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed in ProviderTree for provider: 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 {{(pid=61947) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1318.599221] env[61947]: DEBUG nova.scheduler.client.report [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Inventory has not changed for provider 7263fe00-9b30-4433-9e9f-ec265ed2b8f2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61947) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1319.104644] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.121904] env[61947]: INFO nova.scheduler.client.report [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Deleted allocations for instance ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8 [ 1319.629578] env[61947]: DEBUG oslo_concurrency.lockutils [None req-71fd40dc-3c03-46c1-91cb-7c4a8161396f tempest-ServersTestJSON-1683590871 tempest-ServersTestJSON-1683590871-project-member] Lock "ea067d31-e17d-4fb4-ac04-84fb3a6ca1b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.951s {{(pid=61947) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}